sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/sisob d2.1 r… · web...

81
Project Number: 7PM- 266588 Project Title: SISOB (An Observatorium for Science in Society based in Social Models) Deliverable Type: Report Deliverable Number: D 2.1 WP: 2 Version: 1 Date: April 15, 2011 Title of Deliverable: Literature Review Editor: Richard Walker Authors: Richard Walker (FrontiersIn), Judith Alexander (FrontiersIn), Beatriz Barros (UMA), Eduardo Guzman (UMA), Melissa Cochrane (FrontiersIn), Aldo Geuna (FR), Ulrich Hoppe (UDE), Raimondo Iemma (FR), Soos Sandor (MTA KSZI), Sam Zeni (UDE), José del-Campo, (UMA), Gonzalo Ramos (UMA) Dissemination level: PU Keywords: Literature review, conceptual model, social network, social network analysis, social network indicators, data mining, text mining, mobility, knowledge sharing, peer review, open peer review, old boy network, cronyism Abstract: This report provides an overview of the literature in the main areas of research relevant to the SISOB project. In each case, the review is informed by the basic conceptual model underlying the project. The paper provides a preliminary outline of the model, going on to review the concepts and conceptual tools that will drive the project: social networks, Social Network Analysis and related tools and technologies. The final section provides a review of the concepts, theories and experimental work that underlie the three SISOB case studies, dedicated to researcher mobility, knowledge sharing and peer review respectively, and

Upload: phungminh

Post on 08-Feb-2018

229 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

Project Number: 7PM- 266588

Project Title:SISOB

(An Observatorium for Science in Society based in Social Models)

Deliverable Type: Report

Deliverable Number: D 2.1

WP: 2

Version: 1

Date: April 15, 2011

Title of Deliverable: Literature Review

Editor: Richard Walker

Authors:

Richard Walker (FrontiersIn), Judith Alexander (FrontiersIn), Beatriz Barros (UMA), Eduardo Guzman (UMA), Melissa Cochrane (FrontiersIn), Aldo Geuna (FR), Ulrich Hoppe (UDE), Raimondo Iemma (FR), Soos Sandor (MTA KSZI), Sam Zeni (UDE), José del-Campo, (UMA), Gonzalo Ramos (UMA)

Dissemination level: PU

Keywords:

Literature review, conceptual model, social network, social network analysis, social network indicators, data mining, text mining, mobility, knowledge sharing, peer review, open peer review, old boy network, cronyism

Abstract:

This report provides an overview of the literature in the main areas of research relevant to the SISOB project. In each case, the review is informed by the basic conceptual model underlying the project. The paper provides a preliminary outline of the model, going on to review the concepts and conceptual tools that will drive the project: social networks, Social Network Analysis and related tools and technologies. The final section provides a review of the concepts, theories and experimental work that underlie the three SISOB case studies, dedicated to researcher mobility, knowledge sharing and peer review respectively, and illustrates how the research described relates to the SISOB conceptual model.

Capacities. Science in Society. Collaborative Project

Page 2: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

Copyright © SISOB Consortium 2011-2013 Page ii /lix

Page 3: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

Version history

Version Date Description

1 28.03.2011 Preliminary version for revision by partners

2 14.04.2011 Pre-Final version

3 20.04.2011 Final version

Copyright © SISOB Consortium 2011-2013 Page iii /lix

Page 4: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

Executive Summary

The social impact of science depends on a broad range of factors, some linked to the way scientific knowledge is produced, some to the way it is distributed to actors outside the knowledge production system, and some to the way it is received, exploited and consumed. All involve social interactions among different actors (individuals, groups, institutions) working in different contexts or settings. These interactions constitute a complex social system whose dynamics unfold on a global scale over long periods of time.

The SISOB case studies will focus on just three aspects of the system – mobility, knowledge sharing and the role of the peer review system. However, the goal is to develop concepts and tools that can also be applied outside the specific areas of research covered by the case studies.

The SISOB conceptual model conceptualizes the chain of events leading from discovery to impact as follows

Actors and Knowledge Production Networks produce Artifacts in a Context that affects what they produce and the efficiency with which they produce it;

Distribution Actors and Distribution Networks distribute these artifacts to knowledge users;

Knowledge users use the knowledge, directly or indirectly producing outcomes.

The general goal of SISOB is thus to develop measurements of production networks and/or production contexts and/or distribution networks and/or distribution contexts and to relate these measurements to outcomes.

The knowledge production and knowledge distribution networks cited in the SISOB conceptual model are examples of social networks, a subject of social science research since the pioneering work of Moreno in the 1930s. This and other related work contributed to the birth of a new academic discipline: Social Network Analysis

Later in the 1960s, studies of citation networks made social networks a popular theme in scientometrics. Following up on this work, Crane used social network concepts to study informal pathways for the diffusion of knowledge through scientific communities. More recently, many web-based communities have become social networks in their own right. These developments have inspired scholars to apply computerized Social Network Analysis to scientific communities.

The goal of this report is to provide an overview of relevant concepts, techniques and tools and to relate them to the case studies. It begins by reviewing the main concepts and indicators used to analyze generic social networks and to collect the necessary data and goes on to present taxonomy of the software tools used in this work. These include tools to explore and visualize social networks, tools to extract data from the web (and other sources) and tools to extract information from this data (Knowledge Discovery from Databases)

Copyright © SISOB Consortium 2011-2013 Page iv /lix

Page 5: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

It concludes with a review of literature relevant for the case studies. The main results are summarized below

1. Researcher Mobility is an important element in the formation of knowledge production and knowledge distribution networks. Movement of scientists and scientific knowledge between academic institutions, universities and society, and between different scientific fields is widely believed to be vital to further scientific quality and research development and mobility of academics has thus become a major policy goal for the European Union. The report provides a detailed review of the experimental and observational evidence underlying these assumptions and policies.

2. Knowledge sharing plays a key role in knowledge creation and knowledge distribution networks and can be studied using tools from Social Network Analysis. Supporters of the concept of “Mode 2 science” argue that knowledge production is a transdisciplinary enterprise involving heterogeneous groups of actors. This means that to understand the production of knowledge we need to take account of the context in which it is produced. In this setting, scientific results and products can be seen as boundary objects - abstract or concrete objects that are vague enough to allow collaborating actors from different social worlds to interpret them from their own perspectives but robust enough to keep their own identity despite these differences in interpretation. As such, they are important not only to scientists but also to investors, political actors, journalists etc. Against this background the report reviews, concepts and techniques for the role of these objects and the role of social networks in their generation and transmission.

3. Peer review can be defined as a system of formal evaluation in which scientific research is subjected to the scrutiny of others who are experts in the relevant field. It is commonly used to evaluate scientific papers, contributions to conferences, requests for funding, and on-going projects and, sometimes, labs, departments, and entire universities. As a result, it plays a hugely important role in determining what science is published and funded. In terms of the SISOB conceptual model, peer reviewers are brokers controlling access to knowledge distribution networks. The report summarizes the history of peer review and its alleged failing and biases. These include high cost, lack of transparency in the choice of editors and reviewers, cognitive biases, sexism, nationalism, “institutionalism”, nepotism and so-called cognitive cronyism. Two final sections look at attempts to reform the system (e.g. proposals for open review) and examine how SISOB can analyze peer review with the tools of Social Network Analysis.

Copyright © SISOB Consortium 2011-2013 Page v /lix

Page 6: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

Table of contents

1 Objectives and structure of this document...............................................................1

2 The social impact of science and the SISOB conceptual model..................................2

3 Network science as a new source of insight..............................................................43.1 Social networks, semantic networks etc........................................................................4

3.1.1 Introduction................................................................................................................................43.1.2 Semantic Social Networks..........................................................................................................43.1.3 Network Text Analysis................................................................................................................53.1.4 Pattern based Logical Approaches..............................................................................................6

3.2 Network analysis..........................................................................................................63.2.1 Introduction................................................................................................................................63.2.2 Levels of analysis and measurements.........................................................................................6

3.3 Methods, Tools and technologies................................................................................113.3.1 Methods for identifying social networks..................................................................................113.3.2 Software for Social Network Analysis.......................................................................................123.3.3 Graphical Techniques for Exploring Social Networks................................................................153.3.4 Web Data Extraction Tools.......................................................................................................153.3.5 Data mining – general concepts...............................................................................................17

4 Applying network analysis to science – the SISOB case studies...............................204.1 Network Analysis and Researcher Mobility.................................................................20

4.1.1 Introduction..............................................................................................................................204.1.2 Spillover effects in science........................................................................................................204.1.3 Role of prestige........................................................................................................................214.1.4 Mobility between Academia and Industry................................................................................224.1.5 Obstacles to mobility................................................................................................................22

4.2 Network Analysis and Knowledge Sharing...................................................................234.3 Network Analysis and peer review..............................................................................24

4.3.1 Introduction..............................................................................................................................244.3.2 The history of peer review........................................................................................................254.3.3 Strengths of current models of peer review.............................................................................254.3.4 Weaknesses in current models of peer review.........................................................................264.3.5 Reforming the peer review process..........................................................................................304.3.6 Peer review and SISOB.............................................................................................................31

REFERENCES.................................................................................................................33

Appendix A: Common Network Indicators...................................................................41Introduction.........................................................................................................................41Basic network indicators......................................................................................................41

Indicators for the network (subnetwork) level......................................................................................41Indicators of overall structure...............................................................................................................42Indicators for the node/edge level........................................................................................................42

Indicators of community structure........................................................................................43Subgraph statistics................................................................................................................................43Structural equivalence statistics............................................................................................................43

Copyright © SISOB Consortium 2011-2013 Page /

Page 7: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

Combined indicators............................................................................................................43Correlates of positional features...........................................................................................................43Tie or attachment types and diversity...................................................................................................44Network diversity measures..................................................................................................................44Community profiles...............................................................................................................................44

Appendix B: Software tools for KDD..............................................................................45Open Source.........................................................................................................................45Commercial (DM, KDD and BI – business intelligence --).......................................................47Other software relevant to SISOB (Web Reviews).................................................................48

Copyright © SISOB Consortium 2011-2013 Page /

Page 8: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

1 Objectives and structure of this document

This report is the first deliverable from WP 2 (Conceptual Model). The goal of the report is to provide an overview of the literature in the main areas of research relevant to the project. Of necessity, the literature reviews for each area are informed by the basic conceptual model underlying the project. Chapter 2 thus offers a preliminary outline of the model. Chapter 3 provides a general introduction to the concepts and conceptual tools that will drive the project. Finally chapter 4 provides a review of the concepts, theories and experimental work that underlie the three SISOB case studies, dedicated to researcher mobility, knowledge sharing and peer review respectively. In each case, the review will illustrate how the research described in the literature relates to the SISOB conceptual model.

Copyright © SISOB Consortium 2011-2013 Page 1/59

Page 9: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

2 The social impact of science and the SISOB conceptual modelThe social impact of science depends on a broad range of factors, some linked to the way scientific knowledge is produced, some to the way it is distributed to actors outside the knowledge production system, and some to the way it is received, exploited and consumed. All involve social interactions among actors (individuals, groups, institutions) working in different contexts or settings. These interactions constitute a complex social system whose dynamics unfold on a global scale over long periods of time.

The SISOB case studies will focus on just three aspects of the system – mobility, knowledge sharing and the role of the peer review system. However, the goal is to develop concepts and tools that can also be applied outside the specific areas of research covered by the case studies. The first step is to create a common vocabulary and a common conceptual framework. SISOB thus proposes a shared conceptual model that will inform the research on all the case studies. At the time of writing, work on the details is still in progress. However, it is already possible describe of some of the key entities in the model.

Production Actors. Production Actors are the agents that produce scientific/technological knowledge. An actor can be an individual, an institution or a production network (see below).

Production network. A production network is a network of production actors who collaborate to produce scientific/technological knowledge (e.g. the members of a lab, the members of an institution, the members of a collaboration, a co-authoring network).

Artifact. An artifact is a measurable output produced by an actor (e.g. a scientific paper, a patent, a prototype, a therapeutic protocol etc.).

Production context. A production context consists of a set of conditions, external to the production actors, that influence the kind of artifacts actors produce and the efficiency with which they produce them (e.g. availability of financial resources, know-how, incentives for innovative behavior, career structures etc.).

Distribution Actors. Distribution actors are agents (e.g. reviewers, writers of review articles, journalists) involved in filtering and distributing scientific/technological knowledge to knowledge users (see below).  By their actions, distribution actors can facilitate or hinder the distribution process.

Distribution network. A distribution network is a network of distribution actors (e.g. journalists who cite each other's work, reviewers who often work together)

Distribution context. A distribution context is a set of conditions, external to the distribution actors, that influences the way scientific/technological knowledge

Copyright © SISOB Consortium 2011-2013 Page 2/59

Page 10: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

is distributed to knowledge users (e.g. interest in science, scientific literacy, etc.)

Knowledge users. Knowledge Users use scientific/technological knowledge (e.g. as the basis for new products, new techniques of production, new medical techniques)

Outcomes. An outcome is an effect of new scientific/technological knowledge likely to be of interest to policy makers (e.g. improved scientific knowledge, new products/services etc., employment, improvements in public health, prestige etc.)

The model conceptualizes the chain of events leading from discovery to impact as follows:

Actors and Knowledge Production Networks produce Artifacts in a Context that affects what they produce and the efficiency with which they produce it;

Distribution Actors and Distribution Networks distribute these artifacts to knowledge users;

Knowledge users use the knowledge, directly or indirectly producing outcomes.

Given this basic model we can redefine the goals of SISOB as follows.

The goal of SISOB is to develop measurements of production networks and/or production contexts and/or distribution networks and/or distribution contexts and to relate these measurements to outcomes.

Copyright © SISOB Consortium 2011-2013 Page 3/59

Page 11: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

3 Network science as a new source of insight

3.1 Social networks, semantic networks etc.

3.1.1 Introduction 

The knowledge production and knowledge distribution networks cited in the SISOB conceptual model are examples of social networks, a subject of social science research since the pioneering work of Moreno [1] in the 1930s. This and other related work contributed to the birth of a new academic discipline: Social Network Analysis.

In the 1960s, studies of citation networks [2] made social networks a popular theme in scientometrics. Following up on this work, Crane [3] used social network concepts to study informal pathways for the diffusion of knowledge through scientific communities. More recently, many web-based communities have become social networks in their own right. For instance, the Twitter community uses follower / following relations to create a social network. Facebook structures information spaces through the use of ego-centered networks. These developments have inspired scholars to apply modern techniques of Social Network Analysis to scientific communities (i.e. to networks of actors engaged in the production and distribution of scientific knowledge). Mendeley [4], for example, combines citation networks and reference management software to investigate personal relations between researchers.

3.1.2 Semantic Social Networks 

Mature Social Network Analysis (SNA) [see 5] uses standard mathematical definitions and techniques to characterize so-called two-mode networks or “affiliation networks”. These are essentially bipartite graphs with two types of nodes, known as actors and affiliations. 

Copyright © SISOB Consortium 2011-2013 Page 4/59

Page 12: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

FIGURE 1: TRIPARTITE ACTOR TOPIC MODEL [6]

Typical examples include, e.g., network relating movie actors (actors) to the films in which they have played (affiliations), and networks relating employees (actors) to the companies that employ them (affiliations) (see paragraph 3.2). Another example is the relation between people who make posts to blogs (actors) and the threads to which they post (affiliations).

Malzahn et al., [7] and Harrer et al., [6] extend bipartite social networks with a knowledge dimension represented as an ontology. Peter Mika [8] begins with semantic descriptions of social/personal relations and induces relations based on shared personal interests (affiliations). In both cases, the goal is to integrate semantic and social relationships. Networks constructed in this way can be used to identify people who are interested in similar topics (which they may describe in different terms) and to uncover latent relationships between concepts and instances.

Researchers have proposed extensions to standards to accommodate the needs of this new approach. Michael Galla [9], for example, extends the Web Ontology Language OWL with social relationships and Peter Mika [10] extends W3C’s RDF standard with a social dimension. Although the authors are unaware of any Web 2.0 application that applies this kind of technology to scientific communities, it would be easy to apply in services such as researchGATE [11].

3.1.3 Network Text Analysis 

When the SISOB conceptual model is applied to the case studies, it is often necessary to identify actors who share common interests. This for instance could be a symptom of the kind of cognitive cronyism that has been alleged to impede effective peer review

Copyright © SISOB Consortium 2011-2013 Page 5/59

Page 13: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

(see paragraph 4.3.4). One way of identifying such shared interests is through Network Text Analysis.

This discipline originally emerged from the field of machine learning and content analysis [12] which researchers enhanced with a network dimension [13]. Text analysis identifies networks by measuring co-occurrences of concepts within a moving time window. Concepts are generalized at a low level by stemming and use of thesauri. A meta-matrix then assigns them to higher-level classes such as People, Knowledge/Resources, Events/Tasks, and Groups/Organizations.

Dynamic Text Analysis [14] is a dynamic extension to Network Text Analysis that takes account of changes in the meta-matrix. This technique can be used to track and show changes in the structure of scientific communities e.g. by comparing publication records for different years [15].

3.1.4 Pattern based Logical Approaches 

Another way of identifying similarities among actors is to generate logic based inferences on top of lower level analyses [16]. For instance, it is possible to identify a troll in a blog from the logical pattern of always starting threads and never answering posts [17]. This is of course a very basic example. More advanced applications of logical patterns could make it possible to generate dynamic indicators representing changes in the structure of scientific communities.

3.2 Network analysis

3.2.1 Introduction

Social network analysis (SNA) is a rapidly progressing branch of research, which is developing on the interface between several different fields of the natural and social sciences. In a recent paper in Social Networks, the core journal of SNA, Leydesdorff and Shank [18] demonstrates that papers in the journal are cited by papers in a broad range of disciplines including sociology, physics, computer science and mathematics. However, the same study shows that Social Networks is primarily a sociology journal. This conclusion suggests the need to distinguish between network analysis (NA), in the sense of a toolbox of mathematical techniques for analyzing networks, and Social Network Analysis: the application of these techniques to the study of social relations.

In what follows we provide a brief overview of key methodological issues in NA, and its application to the analysis of social networks (SNA).

Copyright © SISOB Consortium 2011-2013 Page 6/59

Page 14: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

3.2.2 Levels of analysis and measurements

The most general questions in network analysis address the structural, graph-theoretic properties of the network. In the literature, such properties are characterized on three levels of aggregation.

3.2.2.1 Node-level measures

Most measures in NA focus on the structural features of individual elements (nodes) of the graph. Some of the most popular are concepts of node centrality that quantify the connectedness and relative position of elements. The best-known measures are degree centrality, betweenness centrality, closeness- and Eigen-centrality (in the order of observable popularity).

3.2.2.2 Global network-level measures

In many cases, structural features of the whole network are encapsulated in quantitative indices. Genuine network-level measures rely on the edge set of the whole graph (network density, number of components, isolates, path lengths). Other measures provide statistical descriptions of node-level properties characteristic of the network (degree distribution, network centrality, average path lengths etc.).

3.2.2.3 Local cluster-level measures

Several algorithms (n-cliques or clans, k-cores etc.) make it possible to detect and characterize cohesive subgroups (subgraphs, groups, modules etc.) in a larger network. Such structures are often described in terms of subgraph-level measures. Studies of sub-structures in complex networks use a range of topological measures. The clustering coefficient, for instance, expresses the group-level structural of nodes.

3.2.2.4 Positions and block models

Whereas cohesive subgroups are characterized by their local, internal connectivity, positions or blocks are subsets of nodes (actors) are characterized by the similarity of their relations to other actors or categories. This kind of positional analysis often uses so-called block models. In these models, clusters of actors are located in different positions in an image matrix.

Another important notion is the concept of “brokers” [19]. Brokers are nodes in sparsely populated regions of a network that act as a bridge between clusters of densely connected components and control the flow of resources and information between these components. Brokers usually have a high betweenness centrality. In scientific communities, brokers can play a critical role in knowledge sharing. Their positional dimension is strongly related to mobility effects.

Copyright © SISOB Consortium 2011-2013 Page 7/59

Page 15: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

Peer reviewers connecting knowledge production networks to knowledge distribution networks can be seen as brokers (see paragraph 4.3.1)

3.2.2.5 Descriptive vs. predictive analysis

The literature on network models can be divided into descriptive and predictive approaches. Descriptive approaches model and analyze observed relational patterns describing the relevant structural properties of the model in terms of a set of indices and measures (network statistics). Predictive approaches estimate structural features (the probabilities of links between actors in the network, structural indices) of a partially observed network from empirical data or a sample network.

3.2.2.6 Network dynamics and evolving networks

Recent years have seen a change in perspective in the NA-community, with the emphasis gradually shifting from the structure of networks towards changes in the structure of networks, that is, to dynamic aspects of network analysis. Since “real networks are dynamic” (see dynanets.org), changing and evolving through time, realistic models should be capable of coping with these phenomena. This poses a number of challenges.

Time window

Network analyses need to take account of time even when models are cross-sectional, and are not intended to express temporal changes in the domain. To extract a network structures from a real-life domain (be it social communities or scholarly communication) it is necessary to decide the time window to be sampled. For example, a single co-author, citation network, or a document similarity graph may cover publication periods of different lengths in the same analysis. In many cases, the resulting structures are dependent on the window chosen. In rapidly evolving structures, like scientific communities, it may be possible to observe multiple networks with very different densities, distributions and community structures.

Levels of change

Other models may be genuinely dynamic or longitudinal in type. In this case, the goal is to measure changes in the relational structure of a network. This is often achieved by considering a time series of (cross-sectional) networks, representing network states at consecutive time intervals. The extent to which such series can be characterized by classic NA-concepts and measures depends on the components of the network affected by its change:

In the “first-order” case (for example, a longitudinal study of relationships within the same group of people), the set of actors in the network remain intact, while their connections change In such situations, changes can be

Copyright © SISOB Consortium 2011-2013 Page 8/59

Page 16: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

characterized naturally by applying traditional measures to each member of the network series, and operationalizing change as a time series of the resulting values for each measure.

In the “second-order” case, actors and connections vary over time and the size of the network changes over time. In these situations, network-level measures, such as density or the number of connected components can be traced through time as before. However, continuous changes in network membership means that changes in node- and cluster-level properties are difficult to formalize. One of the most demanding challenges is tracking the dynamics of the network’s community structure and the trajectory of individual groups. For example, in a co-word map, where groups represent topics or research directions, each consecutive time slice might be made up from different, or partially overlapping lexicons, each with its own cluster structure. In these conditions, it can be very difficult to identify trends relating initial clusters to subsequent ones and to measure within-cluster changes. For a recent approach see [15].

Roles and positions over time

One technique increasingly used to trace changes over time is positional analysis. Stegbauer and Rausch [20] have applied block modeling with moving time slices to online communities. In bibliometrics, Kronneger, Ferligoj and Doreian [21] have compared longitudinal co-authorship networks for biotechnology, mathematics, physics and sociology for the period 1986-2005. In contrast with Stegbauer and Rausch [20], they use sequences of time slices. Changes are visualized on image matrices. This approach is intermediate between the first-order and second-order cases discussed earlier. While the number of the blocks remains constant over time, the number of actors can change.

3.2.2.7 Types of models: unipartite vs. bipartite networks

Probably the most relevant dimension in practical network analysis is the selection of the type of model best suited for a particular research question. A first step is to decide whether to use a directed or an undirected network, and whether to use weighted or unweighted edges. Directed networks are capable of expressing the directionality of relations (e.g. information flow in citation networks), while undirected models show symmetric relations (e.g. similarity networks of documents). Weighted graphs make it possible to express degrees of relatedness, in cases where this may be useful (e.g. in a network of documents linked by document similarity).

In modeling complex phenomena, there is a clear distinction between unipartite and bipartite graphs. Informally speaking, bipartite graphs depict affiliation networks with two kinds of actors and, implicitly, relations, while unipartite graphs involve one kind of actor and one type of relation. The scheme of a bipartite graph is A–B–C. This

Copyright © SISOB Consortium 2011-2013 Page 9/59

Page 17: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

translates into “A and C is affiliated with B”, with A and C representing one of the kinds, and B the other. A unipartite network would encode the same relationship as A–C, read as “A is in relation with C”. The bipartite model, therefore, says two things, namely, that (1) A and C are connected, and (2) they are connected by virtue of their affiliation to B. The unipartite model only captures relation (1), without exposing the underlying factor B.

As the above outline suggests, bipartite graphs are generally considered to be of great expressive power, making it possible to represent multiple relations in a single model. Project-participant affiliation networks provide a good example[see 22, 23]. In these models the vertices represent a set of projects (type 1 vertices) and a set of participants (type 2 vertices) while the edges connect individual participants to projects (and vice versa). It can be argued that this kind of graph is more effective than unipartite graphs in expressing complex patterns of collaboration between agents. In brief, affiliation networks are extremely useful for modeling multiple relations in complex systems [see the concept of untidy networks in 24]

3.2.2.8 Complex networks and community detection

An important distinction between Network Analysis and traditional graph theory is that the former deals with real-life phenomena, many of which involve large and complex networks with a community structure. In such models, it is possible to identify groups of actors and their associated semantics from the network architecture. An important theme of research in the NA-community is the development of criteria, methods and algorithms to achieve this. The goal is to find coherent sub-graphs in networks representing structural units (e.g. author communities in a co-author network, topics formed out of mutually related terms in a co-word map etc.).

Partitions

Most community detection methods are designed to partition networks – that is to cut them into pieces. These algorithms yield pairwise distinct or mutually exclusive subgraphs representing non-overlapping communities (sets of nodes) in which each actor belongs to exactly one community. The general goal is to identify a community structure in which connections between actors are far denser within than between communities. The most popular method is the so-called modularity-maximizing procedure [25], in which modularity measures to what extent different node types (clusters) are separated from each other. Another established family of approaches is based on the eigen-decomposition of the graph (its underlying adjacency matrix and its derivatives), often referred to as its spectral decomposition [A handy review of existing approaches and new directions, in the case of bibliometic networks, can be found in 26].

Copyright © SISOB Consortium 2011-2013 Page 10/59

Page 18: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

Overlapping communities

In real-life situations communities often overlap. Detecting groups with shared members requires an algorithm that is able to assign network actors to more than one subgraph in the model. The most successful solution to date is called clique percolation [27], and is implemented in the free software Cfinder. CP operates by mapping potentially overlapping, cohesive parts of the graph at different levels of integration (k-cliques). This makes it possible to detect communities with different levels of cohesion ranging from loosely connected communities up to highly integrated, dense subgroups of actors. Cfinder further elaborates the picture by generating a next-order network, in which each community identified by the software is identified as a node and overlaps between communities are represented as edges. This approach makes it possible to observe and analyze relations (overlaps) between groups.

Community detection over time

The CP method has also been applied in to the temporal dynamics of communities. Wang & Fleury [28] propose a framework addressing the influence of temporal variation by “regularizing communities by initializing sub-communities corresponding the cores and proportionality”. Palla et al., [29] formulate an extension on the CP method that uses the size of communities and their age as heuristics. Greene et al., [30] develop a model for tracking communities in dynamic social networks based on events. As in some of the cases cited earlier, the choice of time window is critical. This result is confirmed by Chakrabarti et al., [31].

3.3 Methods, Tools and technologies

3.3.1 Methods for identifying social networks

3.3.1.1 Full network Methods

The goal of full network methods is to collect information about each actor's ties with all other actors. Full network data gives a complete picture of relations in the population. Most techniques of network analysis reply on this kind of data.

Full network data is necessary to properly define and measure many of the structural concepts of network analysis and allows very powerful descriptions and analysis of social structures. Unfortunately, it can also be very expensive and difficult to collect. Obtaining data from every member of a population, and having every member rank or rate every other member are challenging tasks for any but the smallest groups. One way of making it more manageable is to ask respondents to identify a limited number of specific individuals with whom they have ties.

Copyright © SISOB Consortium 2011-2013 Page 11/59

Page 19: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

3.3.1.2 Snowball methods

Snowball methods begin with a focal actor or set of actors. Actors are asked to name some or all of the other actors with which they have ties. These actors are then tracked down and asked for their ties. The process continues until no new actors are identified, or until we decide to stop (usually for reasons of time and resources, or because it is only discovering actors who are very marginal to the group we are trying to study).

The snowball method can be particularly helpful for tracking down "special" populations (small sub-sets of people mixed in with large numbers of others). However, it has two major potential limitations. First, it cannot identify actors who are not connected to other actors (i.e. "isolates"). Second, the result depends on where the snowball starts "rolling". This means the method will not necessarily find all the connected individuals in a population. Snowball approaches can be strengthened by taking special care with the selection of the initial nodes.

3.3.1.3 Ego-centric networks (with alter connections)

In many cases, it is not possible (or necessary) to track down full networks beginning with focal nodes (as in the snowball method). An alternative approach is to begin with a selection of focal nodes (egos), and identify the nodes to which they are connected. We then determine which of the nodes identified in the first stage are connected. This approach can be effective in collecting relational data from very large populations, and can be combined with attribute-based approaches. It can also yield information about the network as a whole, though not as much as snowball or census approaches.

Ego-centric networks (ego only)

Egocentric methods focus on the individual, rather than on the network as a whole. By collecting information on the connections among actors connected to each focal ego, we can get a pretty good picture of individuals’ "local" networks or "neighborhoods". Such information is useful for understanding how networks affect individuals, and can give a (incomplete) picture of the network as a whole.

Analyzing Network Data

Network data can be used to calculate properties of network positions, dyads, and networks as a whole. Properties of network positions include the number of edges associated with a node and the extent to which the node is a bridge between other nodes [32]. Dyads can vary in the strength or reciprocity of their ties, the similarity between the nodes (homophily), the content of the nodes, the number of relation types shared (multiplexity), and the number of communication media used (media multiplexity).

Copyright © SISOB Consortium 2011-2013 Page 12/59

Page 20: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

When studying properties of networks as a whole, researchers can look at such things as the proportion of dyads connected to one another (density), the average path length necessary to connect pairs of nodes, the average tie strength, the extent to which the network is dominated by one central actor (centralization)[32], and the extent to which the network is composed of similar nodes (homogeneity) or of nodes with particular characteristics, (composition). Networks can also be studied in terms of the number of ways they can be divided into subgraphs. For example, networks may consist of multiple components: sets of nodes that are tied directly or indirectly to one another but are not tied directly to nodes in other components. They may also include cliques, in which every node is tied directly to every other node.

3.3.2 Software for Social Network Analysis

Table 1 [33] presents a list of software tools frequently used for Social Network Analysis and provides an outline of their key characteristics.

Copyright © SISOB Consortium 2011-2013 Page 13/59

Page 21: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

Copyright © SISOB Consortium 2011-2013 Page 14/59

Page 22: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

TABLE 1: SOCIAL NETWORK ANALYSIS TOOLS [33]

Copyright © SISOB Consortium 2011-2013 Page 15/59

Page 23: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

3.3.3 Graphical Techniques for Exploring Social Networks

There are two main approaches to constructing graphical images of social networks [34]. The first is based in a search algorithm called Multi-Dimensional Scaling, (MDS). MDS requires that the investigator specify a desired dimensionality - typically, one, two or three. It then uses a search procedure to find optimal locations at which to place the points. Optimal locations are either (1) those that come closest to reproducing the pattern of the original N-dimensional social proximities contained in the data matrix (metric MDS), or (2) those that come closest to reproducing the order, but not necessarily the exact magnitudes, of the original proximities (non-metric MDS).

The second approach is based on, singular value decomposition (SVD). SVD transforms the N original variables into N new variables, or dimensions. The variable with the most variance is always associated with the first dimension, the variable with the second most variance with the second dimension and so on.

Using MDS or SVD together with graphics display programs like MAGE or MOVIEMOL, researchers can determine whether a data set contains interesting structural features.

Several tools for the analysis of social network data (e.g., Pajek and Visone) incorporate one or more algorithms for the visualization and visual exploration of social network data. Most of these algorithms are in the MDS category.

Recently there have been several attempts to work with alternative visualization techniques, incorporate additional data, and to include temporal development of networks into the visualization. One technique is to render the state of the network at different points in time into single images which then can be viewed and explored like movies [35]. Another is to render the network information at a point in time as a two dimensional image and use the third dimension to show the development over time [36, 37]. Examples for embedding additional information into classical network views and new approaches for spatial arrangement algorithms can be found in the work of Lothar Krempel [7].

3.3.4 Web Data Extraction Tools

The SISOB case studies (see section 4) will make intensive use of data on scientific communities and artifacts, collected from the web. This will require the use of Web Data Extraction tools).

Current tools are divided into two categories: tools with a query search interface allowing them to search the “deep web” (e.g. DBMS-based sites such as Amazon) and tools that carry out "traditional" data extraction.

Copyright © SISOB Consortium 2011-2013 Page 16/59

Page 24: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

There are many techniques for developing Web Data Extraction Tools, some of which are summarized in Figure 2 [38].

FIGURE 2: TECHNIQUES FOR DATA EXTRACTION[38].

Figure 3 provides taxonomy of some of the most common tools

  :

Copyright © SISOB Consortium 2011-2013 Page 17/59

Page 25: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

FIGURE 3: CLASSIFICATION OF TOOLS [38]

3.3.4.1 Data extraction tools specialized in SNA

Some Social Network Analysis tools include data capture capabilities relevant to SNA. One such tool is Netminer. This is a commercial tool whose data capture features are provided by external modules. The software makes it possible to define association rules for mining relational patterns from existing data bases and provides additional analysis features.

www.netminer.com

Another commercial tool is Commetrix. Commetrix extracts data from mails, newsgroups, and discussion boards. The extraction process explicitly supports time events to extract data for dynamic analysis. Analysis and visualization components are built into the tool.

www.commetrix.de

Another tool which can extract data from web is NodeXL [39]. NodeXL is free software that can extract data from Twitter, flickr, and Youtube by using the APIs of these websites. It also provides analysis and visualization features.

http://nodexl.codeplex.com

The Data-Multiplexer-Demultiplexer (DMD) is a research prototype developed by COLLIDE research group [37] which may well be suitable for use in SISOB. DMD can extract data from mailing lists, Bibtex bibliographies, discussion boards, and wikis as well as transform them to common formats for tools like Pajek and UCINET.

3.3.5 Data mining – general concepts

Knowledge Discovery in Databases (KDD) is the process of identifying valid, novel, useful, and understandable patterns from large datasets. At the heart of KDD is Data Mining (DM) - the use of pattern recognition and statistical techniques to discover meaningful new correlations, patterns and trends in large amounts of data stored in repositories [40, 41].

Copyright © SISOB Consortium 2011-2013 Page 18/59

Page 26: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

There are many different approaches in the KDD process. Maimon [40] summarizes the main issues. The knowledge discovery process is iterative and interactive, consisting of the nine steps illustrated in Figure 4. Note that the process is iterative at each step, meaning that it may be necessary to move back and forward between steps.

FIGURE 4: THE KDD PROCESS

3.3.5.1 STANDARDS

Data mining and statistical models generated by commercial data mining applications are often used as components systems for Customer Relationship Management (CRM), Enterprise Resource Planning (ERP), risk management, and intrusion detection. In the research community, data mining is used in systems processing scientific and engineering data. Data mining standards make it much easier to integrate, update and maintain such systems [42, 43].

3.3.5.2 MODELS

There are two main standards for defining the models generated by data mining techniques. They are PMML and CWM-DM. The most popular is PMML. Below we summarize key features of each standard.

Copyright © SISOB Consortium 2011-2013 Page 19/59

Page 27: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

PMML (Predictive Model Markup Language) [Pecther-2009][Guazelli-2009b]

http://www.dmg.org XML language for describing statistical and data mining models in an application

and system independent fashion. With PMML, models can be exchanged easily between systems and persisted to repositories.

Version 4.0 was released in May 2009 (currently v.4.0.1) Is being used to do extensions: An Extended Predictive Model Markup Language

for Data Mining (EPMML) [Zhu-2010] Increasing number of applications that support PMML (see table 4.4.2) There exists tools to validate and convert models:

o Zementi's PMML converter (http://www.zementis.com/pmml.htm). You can use the to validate your PMML file against the specification for versions 2.0, 2.1, 3.0, 3.1, 3.2, and 4.0. If validation is not successful, the converter will give you a file back with explanations for why the validation failed

CWM-DM (Common Warehouse Metamodel)[Poole-2003]

http://www.omg.org/technology/documents/modeling_spec_catalog.htm Open industry standard defining a common metamodel and eXtensible Markup

Language (XML) based interchange format for meta data in the data warehousing and business analysis domains

Version 1.1 was released in April 2003 (currently v.1.1) [OMG-2003]

3.3.5.3 TOOLS for Statistics, Machine Learning, Data Mining or Knowledge Discovery in Databases

There are many tools and libraries for KDD. These can be grouped into two main categories: open source and commercial. Appendix B provides a list and a brief description of these tools.

Copyright © SISOB Consortium 2011-2013 Page 20/59

Page 28: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

4 Applying network analysis to science – the SISOB case studies

4.1 Network Analysis and Researcher Mobility

4.1.1 Introduction

Mobility is an important element in the formation and maintenance of knowledge production and knowledge distribution networks. Movement of scientists and scientific knowledge, and hence social capital, between different academic institutions, university and society, and between different scientific fields is believed to be vital to further scientific quality and research development. Mobility of academics has thus become a major policy goal for the European Union.

It is generally believed that mobility of scientists facilitates knowledge and technology transfer, creation of networks and productivity [44-46]. These assumed positive effects are related to the embedded character of scientists’ human and social capital [47, 48]. When they move, scientists increase their own value as human capital [49-54] and contribute to the receiving institution’s stock of social capital [55-57]. In this perspective, it is assumed that mobility benefits both the research system and individual researchers.

The benefits of mobility for academic careers are largely due to enhancement of social capital. Mobile researchers gain access to academic networks, develop scientific contacts and widen their communication channels. Mobile researchers, moreover, receive intellectual stimuli from their new environment and enhance their personal skills. Early studies in the US suggested that mobile academics have higher levels of scholarly productivity than academics employed by their PhD awarding institution [58].

Many authors have suggested that in the US, mobility is strongly associated with scientific merit and encouraged by universities. By contrast, most countries in Europe are characterized by academic inbreeding and a reluctance of academics to move [59].

4.1.2 Spillover effects in science

Mobile researchers have been shown to create positive spillovers by enabling knowledge flows and exchange of expertise. Studies on social capital of mobile inventors have shown that links to the original location are maintained and that knowledge flows are deeply embedded in labor mobility [60-62]. Similarly, Azoulay et al., [63] find that researchers moving to a new institution increase their citations from the destination university noticeably, while citation rates from the origin institution

Copyright © SISOB Consortium 2011-2013 Page 21/59

Page 29: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

are not affected. The evidence shows that researchers can increase their visibility and credibility by moving to a different academic environment.

A move to a new environment is likely to have a positive effect both on the productivity of a researcher and on that of the receiving department. Literature on mobile inventors has indeed found a positive link between inventor mobility and productivity [64]. On the other hand, there is very little evidence on the effect of mobility on research productivity. Some weak evidence suggests that immobility has a negative impact [58] and that post-doctoral research abroad may be especially important [65]. However, other evidence indicates that mobility may be an intrinsic characteristic of productive researchers rather than a direct contributor to productivity [66]. Some studies have shown that researchers tend to adjust their productivity to that of the department where they work [67]. However, Kim et al. [68] have demonstrated that this effect was less important in the 1990s., than in earlier periods. Nonetheless, top scientists still tend to agglomerate in high ranked universities. Looking at the mobility and promotion patterns of a sample of 1,000 top economists, Coupé et al. [69] suggest that the sensitivity of promotion and mobility to production diminishes with experience, indicating the presence of a learning process.

Cañibano et al. [70] argue that it is the qualitative dimension of mobility impact that is most important. Most internationally mobile researchers are embedded in larger networks, co-operating with foreign researchers and gaining access to international sources of funding.

The positive effects on their new institutions have become more evident in recent years. In the UK, university departments have used tactical hiring as a mechanism to increase their credibility. In the case of short-term research visits, receiving institutions hope to indirectly increase their reputation in the home environment of the visiting scientist.

4.1.3 Role of prestige

Though mobility has become an important element in US and UK academic careers, it is driven by reputation and largely limited to a small group of elite universities [71]. Several papers have considered the relationship between the prestige of PhD granting and hiring institutions and researchers’ early careers. Most find that for young scientists looking for their first position, the prestige of the university where they obtained their doctorate is more important than their productivity during their PhD training [72-75]. Crane [72, 73] studied the probability of young scholars being hired by one of the top 20 departments in their field and found that the best predictor was the prestige of their doctoral program. A study of neural network researchers by Debackere and Rappa [76] suggests that prestige matters most early on in a scientist's career, but that at later stages it no longer plays a significant role. Chan et al. [77] find that scientists coming from highly ranked PhD granting institutions have a better chance than others of finding a position in such an institution. A scientist’s

Copyright © SISOB Consortium 2011-2013 Page 22/59

Page 30: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

choice of where to take her Ph.D thus involves tactics of professional socialization with the right choice providing access to elite networks and career advantages [78].

At later stages, the impact of institutional prestige declines and productivity becomes at least equally important [74]. Nonetheless, academia’s prestige culture still favors elite institutions. Chan et al. [77] find that very few researchers are able to move from a lower to a higher ranked institution and that these few exceptional scientists are two times more productive than the average academic at the destination university. In a recent paper Kim et al., [68] find that in the US prestige culture and its agglomeration effects remain very strong.

4.1.4 Mobility between Academia and Industry

A special case of researcher mobility is the transition between the private and public sector. Intersector job changes are encouraged by policy makers. Dietz and Bozeman [79] have found the career patterns of academics coming from industry are very different than that of their peers.

Zucker et al. [80] have studied the probability of an academic star moving into industry and Crespi et al. [81] have investigated the probability of similar moves by inventors. Both these studies show the relevance of embedded human and social capital for knowledge and technology transfer and identify key conditions that influence academia to industry and firm-to-firm mobility.

4.1.5 Obstacles to mobility

One of the main obstacles to the mobility of scientists is academia’s elite culture. Given that researchers aim to enter the best departments and that departments try to hire the most promising staff, choice and opportunities are limited [71]. In the US and the UK, selection starts at an early stage. The best students enroll in the best undergraduate programs, gain the best degrees and in turn have the best opportunities to choose the best PhD program [71]. In continental Europe, mobility is particularly low due to tenured academic staff’s civil servant status and lifelong appointments. Different organizations tend to be very similar and competition among them is weak. As a result, a move is not seen as a possibility for advancement [59]. Opportunities for promotion depend to a large extent on social ties, further limiting opportunities for mobility [82, 83].

Intersector mobility poses a different set of problems. Marcson [84], Krohn [85], Kornhauser [86] and Hagstrom [87] have analyzed the ‘role strain’ involved in job transitions between academic and business environments. The focus of these studies is on difficulties in adaptation to new norms and patterns of behavior. In general it appears that researchers in Europe have far fewer opportunities than their US peers. The R&D labor market is not very strong [88]. While a transition to industry is possible, a return to academia is very difficult.

Copyright © SISOB Consortium 2011-2013 Page 23/59

Page 31: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

A short essay by Melin [89] points out an additional obstacle to mobility. In interviews with Swedish postdoctoral returners he found that 10-20% were unable to exploit the knowledge they had gained abroad and that many did not receive the recognition they had expected. This failure in knowledge transfer may be an intrinsic characteristic of Europe’s highly inflexible academic market.

4.2 Network Analysis and Knowledge SharingKnowledge sharing, including knowledge transfer and knowledge creation in and between teams and communities, plays a key role in knowledge creation and knowledge distribution networks and can thus be studied using tools from Social Network Analysis. Valente [90] provides an overview of existing work. One approach is to trace the diffusion of knowledge within networks using methods similar to those Rogers [91] used to trace the diffusion of innovations. For example, these methods make it possible to identify early adopters with an influential position in the network.

Recent research on knowledge creation and sharing has introduced other innovative concepts. One important approach centers on the concept of “Mode 2 science” [92]. In this view, knowledge production is a transdisciplinary enterprise involving heterogeneous groups of actors. This means that to understand the production of knowledge we need to take account of the context in which it is produced and in the particular the market for knowledge.

In this approach scientific results and products are boundary objects [93] - abstract or concrete objects that are vague enough to allow collaborating actors from different social worlds to interpret them from their own perspectives but robust enough to keep their own identity despite these differences in interpretation. As such they are important not only to scientists but also to investors, political actors, journalists etc.

In this setting, transdisciplinarity may be interpreted as multidisciplinarity within a broader field. One example of a transdisciplinary field is nanotechnology, which integrates contributions from physics, chemistry, and electrical engineering, as well as from a broad range of application areas [94]. It is obvious that this kind of field involves significant transfers, integration and re-definition of knowledge. From a methodological point of view, disciplinary heterogeneity can be addressed by mapping different networks of collaboration and cooperation, and analyzing them individually and comparatively. Heimeriks et al. [95] have demonstrated the effectiveness of this approach in biotechnology research.

Another factor that needs to be considered is the role of context in the knowledge creation process. The tacit dimension of knowledge and the influence of the context in which knowledge is externalized have been described by Nonaka and Takeuchi [96]. The recent network research literature suggests that the influence of the context, and especially institutional context, can be observed at the macro level. Wagner and Leydesdorff [97] compare the evolution of co-authorship networks in six distinct fields with a known model for network evolution (“preferential attachment”, as

Copyright © SISOB Consortium 2011-2013 Page 24/59

Page 32: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

described by Albert & Barabasi [98]). They explain the divergences between the observed data and the model in terms of institutional constraints, going on to categorize authors into continuants, transients, newcomers and terminators. These categories, first introduced by Braun et al. [99], help to explain actors’ productivity and patterns of cooperation between them.

Nonaka and Takeuchi’s proposal leads to a model of knowledge and actor heterogeneity which can be addressed by social network analysis [100]. In this setting it is possible to model informal knowledge like “who knows what” and to generate weighting indicators like the level of trust associated with authorities [6]. The analogy with implicit and explicit knowledge leads to models, which make a distinction between individual knowledge and common knowledge. These models allow the generation of affiliation networks in which network properties like holes or measures of individual entities like betweenness or brokerage can be used to identify factors facilitating or hindering knowledge flow.  Recent studies show that there are relations between network structure and knowledge production although these relations are specific to cases and cannot be easily generalized ([101] [102]).

Other recent developments include the work of Van den Besselaar & Heimeriks [103] who show that a method based on combined use of title words and cited references can provide promising two dimensional indicators representing the position of research topics on a network map for a discipline. Another approach is to apply techniques developed for the representation and analysis of semantic networks to social networks. Mika et al. [10] demonstrate that semantic technologies can be used to aggregate and analyze data from different heterogeneous sources. In a study of semantic web research, they confirm the positive effect of structural holes [19] on the innovativeness of researchers. Other insights into knowledge sharing and creation may come from studies of technologies for information exchange and communication. For instance, Bullinger et al. [104] have used qualitative analysis techniques to provide a first tentative taxonomy of Social Research Network Sites (SNRS).

4.3 Network Analysis and peer review

4.3.1 Introduction

Peer review can be defined as a system of formal evaluation in which scientific research is subjected to the scrutiny of others who are experts in the relevant field. It is commonly used to evaluate scientific papers, contributions to conferences, requests for funding, and on-going projects and, in some countries, labs, departments, and even entire universities. As a result, it plays a hugely important role in determining what science is published and funded. In terms of the SISOB conceptual model, peer reviewers act as brokers controlling access to distribution networks. While a positive peer review is no guarantee of success (see below) negative reviews block access to the network. In other words, science that fails to pass peer review is unlikely to have any impact. To the extent that careers are determined by publication and ability to

Copyright © SISOB Consortium 2011-2013 Page 25/59

Page 33: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

attract funding, peer review determines the course of scientists’ careers, and the success of their labs and institutions. In summary, peer review is an essential element in the organization of modern science. As such it has itself become an object of study.

In the period from 1970 to 2000, scholars generated a steadily expanding literature describing its history and failings, characterizing its underlying mechanisms and making suggestions for improvement. Much of this work is ably summarized in major review articles by Campanario [105] and Armstrong [106]. Since 2000 the number of studies of peer review has tended to fall – perhaps because many originally controversial results have won wide acceptance in the community. In what follows we refer to this later literature, wherever useful. However, we find that many recent papers do little more than recapitulate and replicate earlier findings. We also note that the majority of studies focus on reviews of journal papers and that there have been far fewer analyses of peer review by funding agencies. This is a general weakness of the literature.

4.3.2 The history of peer review

Several authors have written about the history of peer review [107-110]. The main facts are not in dispute and can be summarized as follows:

1. Peer review is ancient. Examples have been documented as far back as the 9 th

century C.E, when Ishap bin Ali al Rahwi, a Syrian physician, called on his colleagues to keep notes of the condition of their patients. If the patient later died the notes would be examined a by a committee of physicians which would decide if he/she had received adequate care. Forms of peer review more familiar to the modern scientists are documented as far back as 1732, when, the Royal Society of Edinburgh had instituted a select committee of scientists to decide which papers to publish in its Philosophical Transactions [107].

2. Peer review only attained its current importance in the second half of the twentieth century. Between the mid-19th century when the number of scientific papers produced first began to exceed the ability of journals to publish them, and the mid-20th century when modern practices first began to come into force, publication decisions usually depended on the policies and decisions of the editor. When, in 1905, Max Planck and Wilhelm Wien decided to publish five extraordinary papers by Albert Einstein that were to revolutionize modern physics, they made their decision without any form of external review [111], . The Journal of the American Medical Association (JAMA) did not use outside reviewers until the 1940s [107]. Lancet did not institute systematic peer review until 1976[110].

3. The pervasive application of peer review is associated, on the one hand with technological change (the introduction of the photocopier which made it easy to distribute papers to reviewers), on the other with the exponential growth

Copyright © SISOB Consortium 2011-2013 Page 26/59

Page 34: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

in scientific production that began in the period after the end of World War II [107].

Today, the exponential growth in the production of scientific papers continues but new web technologies creates opportunities for new forms of peer review that conserve the strengths of the common system, while remedying its well-recognized weaknesses.

4.3.3 Strengths of current models of peer review

All of the papers reviewed in this study support the use of some kind of peer review to select papers for publication and (in some cases) projects for funding. The main argument is that peer review improves the quality of scientific publications [112]) Armstrong reports that “journal peer review is commonly believed to reduce the number of errors in published work, to serve readers as a signal of quality and to provide a fair way to allocate journal space” [106]. Surveys of authors and expert reviewers [113-115] confirm this widely held view.

4.3.4 Weaknesses in current models of peer review

Despite widespread belief in the essential role of peer review, and its contribution to quality, all the papers reviewed point to weaknesses in the current model. In many cases, authors use evidence from experimental and observational studies. Below we summarize the main arguments.

4.3.4.1 Time and money

The first, very common objection is that peer review is slow and expensive. Editing a journal is a time-consuming consuming process that absorbs a huge effort from editors and their assistants. Although much of this work is unremunerated or badly paid, it nonetheless gives rise to significant administrative costs. As early as the 1970s the New England Journal of Medicine’s review process required 6-7 person years of effort and more than $100.000 of office expenditure per year [116]. Scanlan [117] reported in the early 1990s that reviews cost biomedical journals an average of $75-150 per article.

Obviously peer review process involves the work not only of editors and their assistants but that of reviewers. Reports of the mean time a reviewer takes to review a paper range from 2 to 6 hours [109]. In some disciplines, well-known reviewers may review as many as 15 manuscripts in a year. Unsurprisingly, editors report that delays in returning reviews are common. Further delays are introduced by requests for revision, which do not always contribute to the quality of the paper (see below). Although we were unable to find a systematic study of times to publication, anecdotal evidence suggests that the usual lag is more than six months and sometime more than

Copyright © SISOB Consortium 2011-2013 Page 27/59

Page 35: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

a year. For many commentators [108, 110], this represents an unacceptable slowing down of the scientific process.

4.3.4.2 Choice of editors and reviewers

Once the editor of a journal sends an article for review, the ensuing process is highly formalized. Many commentators observe, however, that the choice of editors and reviewers is less transparent. Given that editors decide which papers are given out for review this is an important issue. To date, however, there seems to have been relatively little research on the way editors are selected.

A study by Yoels [118] suggests that editors from top American universities tend to recruit editors from the same universities. Other studies (none recent) confirm that at least in some disciplines nearly all reviewers come from a small group of top universities. It is not known however how far this tendency persists today.

A second linked issue is the choice of reviewers. Many observers have suggested that the power to choose reviewers gives editors undue influence over the outcome of reviews [119]. Several studies have suggested that referees tend to come from the same institutions as editors [120]. Another so far unstudied issue is to what extent reviewers and authors form “cliques” with a small group persistently reviewing each other’s papers.

But, even in the absence of obvious sources of bias, there is no guarantee that the reviewers chosen are those most qualified to do the job. Campanario reports that much reviewing work is done, not by senior scientists, but by their more junior colleagues, sometimes by direct decision of the editor, sometimes, as a form of hidden “subcontracting” [109].

4.3.4.3 Cognitive Biases

Studies over many decades have shown that human reasoning is subject to systematic cognitive biases. Peer review process is not exempt from this tendency. Below we describe some of the reported sources of bias.

Preference for positive results

In a classical study, Mahoney [121] asked four groups of reviewers to referee a manuscript. The first group received manuscripts that described positive results, the second group papers with negative results; the third and fourth groups papers with mixed and no results respectively. Experimental procedures were identical in all papers.

The results were surprising. Referees who reviewed papers with positive results gave a higher score to the papers they reviewed than the other referees. Conversely, referees showed a clear bias against papers reporting negative or mixed results. Interestingly many of their criticisms focused on methodology, which, in reality, was

Copyright © SISOB Consortium 2011-2013 Page 28/59

Page 36: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

identical in all the papers. If we accept the idea that scientific progress depends on the falsification of accepted hypotheses [122], this tendency is clearly counter-productive.

Difficulty in publishing replication studies

Surveys of reviews and of journal publications suggest that reviewers are unwilling to approve the allocation of scare space and funds to studies whose main goal is to replicate a previous study [123, 124]. Many journals and funding agencies, including the European Union’s Framework Program [125] include novelty among their selection criteria for papers and articles. The result is that many results with far-reaching theoretical or practical implications have never been replicated in the open literature. Reviewer practices thus create a paradoxical situation in which replicability is considered essential for good science but actual replication is actively discouraged.

Copyright © SISOB Consortium 2011-2013 Page 29/59

Page 37: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

False cues to quality

An extremely useful review article by Armstrong [108] suggests that many reviewers use “false cues” in judging the quality of a paper. His examples include the following:

Use of statistics. A substantial body of experimental and observational studies suggest that studies reporting tests of statistical significance are significantly more likely be published that studies that do not. For instance Atkinson, Furlong and Wampold [126] report an experiment in which reviewers were asked to evaluate three versions of a bogus manuscript that differed only in the reported statistical significance of their results. The papers with non-significant results were rejected three times more often than those whose results were reported as significant. Interestingly reviewers mostly based their rejection decisions on the experimental design, which was actually identical in all the papers.

Obscure writing. An earlier paper by Armstrong [127] reports an experiment in which referees were asked to evaluate a short (fake) text, apparently extracted from the conclusions of a manuscript. In one version the text used long words and complicated sentence structure. The other used simpler vocabulary and less convoluted structures. Unfortunately for the prejudices of this author, the reviewers systematically preferred the more obscure versions of the text, which they apparently interpreted as a sign of “intelligence” and “depth” [108].

Other false cues. Armstrong [108] points to other “false cues” that reviewers may use to judge the quality of a paper, in particular the use of sophisticated statistical methods and large sample sizes. Unfortunately he provides no experimental evidence that these factors actually influence acceptance or rejection of papers.

Confirmatory bias. An important issue in the peer-review process is so-called confirmatory bias: the strong human tendency to give greater weight to evidence that confirms our own views than to evidence that contradicts them. Given the operational difficulty in identifying reviewers’ opinions, there have been relatively few studies of this potential source of bias. Initial evidence is presented in Mahoney [121]. A rare ethnographic study of peer review in a funding agency by Travis [119] confirms that reviewers tend to accept proposals that agree with their own scientific opinions and to reject those that do not.

4.3.4.4 Sexism, Nationalism, “Institutionalism” and Nepotism

Sexism. Lloyd [128] shows that manuscripts with female author names have a far higher acceptance rate when they are reviewed by female rather than male reviewers (62% vs. 21%). A widely quoted study of grant awards in Sweden [129] suggests that proposals from male candidates receive systematically

Copyright © SISOB Consortium 2011-2013 Page 30/59

Page 38: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

higher evaluations than those from female candidates. A recent follow-up study [130] shows no significant change in this tendency.

Nationalism and “institutionalism”: A study by Link [131] suggests that American authors are significantly more likely to accept a paper by another American author than a paper by an author of a different nationality. Peters and Ceci [132] report a quasi-experiment showing that papers by authors from high-prestige institutions have a significantly higher chance of acceptance than similar papers by authors with less prestigious affiliations.

Nepotism. Some of the earliest criticisms of the peer review process claimed that it systematically favors some authors over others – a tendency sometimes referred to as “old-boyism”, “croyism” or “particularism” [119]. Wade [133], cites a widely held opinion that “the peer review system consists of a group of committees whose members hand out grants to each other and to their friends”. The two studies of grant awards in Sweden cited earlier [129, 130] show that grant applicants sharing an affiliation with a member of the selection committee have a significantly higher chance of receiving an award than other applicants of the same gender with the same reported level of scientific productivity.

4.3.4.5 Ability to detect errors, plagiarism and fraud

In addition to the suggestion that peer review is subject to systematic bias, many authors have suggested that it is unable to reliably detect errors, plagiarism and fraud. Armstrong [108] quotes a long series of studies showing that a high proportion of published papers contain serious errors in bibliographical references, in interpretations of the work by other scientists, and in the use of statistics. He cites a study by Epstein (Epstein 1990) showing that in many cases, reviewers do not detect obvious cases of plagiarism. Recent cases of scientific misconduct [134] have shown that it is relatively easy for authors to deliberately mislead reviewers. This has led authoritative commentators to conclude that “clever fraud can only be detected by people working in the lab who have access to the raw data or by other labs who try to replicate (...) later” [135].

4.3.4.6 Reliability of peer review

In view of the many weaknesses described in the previous section, it is not a surprise that many studies cast doubt on the reliability of the peer review process.

Inter-referee agreement

It is a common experience for authors to receive radically different assessments of their work from different reviewers. Experimental studies of the peer review process show very low levels of agreement among referees. For example, in the study by Mahoney [121] cited earlier, there was very little correlation between rankings by different reviewers. Cole, Cole and Simon submitted a series of grant proposals previously reviewed by the NSF to a second panel of independent reviewers. In a

Copyright © SISOB Consortium 2011-2013 Page 31/59

Page 39: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

large number of cases, the second review gave the opposite result to the first (acceptation of a previously rejected proposal, rejection of a previously accepted proposal). Subsequent analysis showed that most of the variance in proposal attributes could be attributed not to the inherent quality of the paper but to differences in opinion among the reviewers. In other words, “getting a research grant depends to a significant extent on chance” [136].

Ability to predict success – inhibition of innovation

Campanario [109] distinguishes errors by reviewers into two classes. Type I errors – where reviewers accept papers that do not merit publication and Type II errors – where they reject papers that should have been published.

Analyses of citations show that a large proportion of all scientific papers receive less than 10 citations and that only very few receive more than 50 [137]. This suggests that type I errors are extremely common. Evidence for Type II errors is obviously harder to come by. Nonetheless, authors of important scientific discoveries and highly cited papers cited in Campanario [109] often report encountered serious obstacles during the review process. In some cases resistance from reviewers was such that very important discoveries (e.g. the First Law of Thermodynamics) were published in very obscure journals. These findings confirm suggestions by Mahoney [121] and later authors that the peer review process can create serious obstacles for innovative thinking and results.

4.3.5 Reforming the peer review process

The alleged weaknesses of the peer review system, have inevitably given rise to proposals for reform, several of which have been adopted by reputable journals and publishers. Below we briefly summarize some of the most popular.

4.3.5.1 Result-blind review

In view of reviewers’ observed “confirmatory bias”, several authors [105, 121] have argued that acceptance or rejection of a study should not depend on the results themselves but exclusively on the quality of the methodology and (possibly) the importance of the findings. This implies that reviewers should be blinded to the results section of the papers they are reviewing. To the knowledge of the authors, no journal has attempted to put this suggestion into practice. Interestingly however, the guidelines for reviewers for Frontiers Open Access journals state that reviewers should reject papers containing factual errors but should not attempt to evaluate authors’ interpretation of their results [138].

4.3.5.2 Author-blind review

Given the evidence for nepotism, sexism, “institutionalism” and nationalism there have been several proposals that reviewers should be blinded to the names of authors

Copyright © SISOB Consortium 2011-2013 Page 32/59

Page 40: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

and their affiliations. Other commentators have observed, however, that reviewers familiar with their field can often identify anonymous authors from their approach, ideas, style and references [112, 139, 140]. A survey of experimental studies [141]suggests that this kind of blinding has little effect on the outcome of reviews.

4.3.5.3 Removing the anonymity of reviewers

Another common suggestion for reforming the review process is to remove the anonymity of reviewers. However, a review of several studies by Benos et al., [110] concludes that that the removal of reviewer anonymity makes little difference to the quality of reviews. Understandably, authors of favorable reviews are reported to be more willing to give up their anonymity than authors of negative reviews. A number of publishers have instituted systems of open review that take this preference into account. For example, the Frontiers open review system, referred to earlier, only publishes the names of reviewers for papers that are accepted for publication. Reviewers of rejected papers remain anonymous. This policy encourages reviewers to take responsibility for the papers they select while allowing them full freedom to express negative criticism.

4.3.5.4 “Open review”

Several online journals (such as Biology Direct; Atmospheric Chemistry and Physics; Electronic Transactions on Artificial Intelligence) are experimenting with systems of “open review” [110]. To the knowledge of the authors, there has been no formal evaluation of the effectiveness of these systems.

Another common practice (especially in the physics and mathematics communities) is to publish preliminary papers on pre-print servers, where other scientists can comment and make suggestions for improvement, prior to formal submissions for publication. Several major publishers have adjusted their rules for publication to accommodate this practice (for example, Nature [142])

An alternative model of open review, introduced by the Frontiers publishing group, is to open papers for community comment after publication.

4.3.6 Peer review and SISOB

The central theme of SISOB is the way social networks influence the social impact of science. The critiques of the peer review system described in the previous paragraphs include numerous examples.

Following Travis et al., [119] we distinguish two ways in which social networks can influence the peer review process.

1. The first is “old boy cronyism”: the tendency of editors to choose reviewers of their own gender and national origin, or coming from the same institution. It is possible (though unproven) that the formalization of review procedures and

Copyright © SISOB Consortium 2011-2013 Page 33/59

Page 41: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

modern rules on conflict of interest have reduced the importance of this “traditional” form of bias.

2. The second is so-called “cognitive cronyism”: the tendency of reviewers to favor authors whose work confirms their own theories and ideas. Travis et al. suggest that this tendency may be especially pernicious in highly innovative areas of science, in which leading researchers disagree about fundamental principles.

Hypotheses about “old boy cronyism” and “cognitive cronyism” can be formulated in terms of hypotheses about the operation of social networks. Such hypotheses can be formulated in the general form:

The probability of outcome x is influenced by membership in network y (1)

A more specific hypothesis might be formulated as follows:

The probability of publishing a paper in journal x is increased by affiliation to institution y (old boy cronyism)

or

The probability of publishing a paper in journal x is increased by membership of a network of authors and reviewers sharing belief y (cognitive cronyism).

Or

An open review process reduces certain kinds of bias

The group that will conduct the SISOB study of peer review has access to an unprecedented corpus of data on papers submitted for publication, the origins of authors and reviewers, the results of the review process, and subsequent post-publication evaluations by readers.

Using results from this literature review, the group will formulate hypotheses concerning the way “old boy” and “cognitive cronyism” biases the review process and test these hypotheses using the data it has collected. It will go on to look at the way other aspects of relevant social networks can influence the peer review system favoring or hindering the publication of high impact work.

Copyright © SISOB Consortium 2011-2013 Page 34/59

Page 42: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

REFERENCES

1. Moreno, J.L., Who Shall Survive? 1934, Washington, DC: Nervous and Mental Disease Publishing Company.

2. de Solla Price, D.J., Networks of Scientific Papers. Science, 1965. 149: p. 510-5.3. Crane, D., Invisible colleges: Diffusion of knowledge in scientific

communities1972, Chicago: University of Chicago Press.4. www.mendeley.com. [cited 2011 April 6th].5. Wasserman, S. and K. Faust, Social Network Analysis, 1994, New York: Press

Syndicate of the Cambridge University Press.6. Harrer, A., et al., Combining social network analysis with semantic relations to

support the evolution of a scientific community, in Mice, Minds, and Society — The Computer Supported Collaborative Learning (CSCL) Conference 2007 , C. Chinn, G. Erkens, and S. Puntambekar, Editors. 2007, International Society of the Learning Sciences. p. 267-276.

7. Krempel, L., Visualisierung komplexer Strukturen - Grundlagen der Darstellung mehrdimensionaler Netzwerke. Schriften des Max-Planck-Instituts für Gesellschaftsforschung, 2005: p. Sonderband Campus, Frankfurt a.M

8. Mika, P., Ontologies are us: A unified model of social networks and semantics. Web Semantics: Science, Services and Agents on the World Wide Web, 2007. 5(1): p. 5-15.

9. Galla, M., Social Relationship Management in Internet-based Communication and Shared Information Spaces, in Computer Science, Ph.D Thesis, 2004, Technische Universität München: Munich.

10. Mika, P., T. Elfring, and P. Groenewegen, Application of semantic technology for social network analysis in the sciences. Scientometrics, 2006. 68(1): p. 3-27.

11. www.researchgate.net. [cited 2011 April 6th].12. Carley, K.M., Coding Choices for Textual Analysis: A Comparison of Content

Analysis and Map Analysis, in Sociological Methodology, C.W. Roberts and N.J. Mahwah, Editors. 1993, Blackwell Publishers: Cambridge, MA. p. 75-126.

13. Diesner, J. and K.M. Carley, Revealing Social Structure from Texts: Meta-Matrix Text Analysis as a novel method for Network Text Analysis , in Causal Mapping for Information Systems and Technology Research: Approaches, Advances, and Illustrations, V.K. Narayanan, Editor, 2004, Idea Group Publishing: Harrisburg, PA.

14. Carley, K.M., Dynamic network analysis, in Dynamic Social Network Modeling and Analysis: Workshop Summary and Papers, K. Breiger, K.M. Carley, and P. Pattison, Editors. 2003: Committee on Human Factors, National Research Council. p. 133-145.

Copyright © SISOB Consortium 2011-2013 Page 35/59

Page 43: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

15. Kampis, G., et al. Dynamic Social Networks and the Textrend / CIShell Framework. in Applied Social Network Analysis Conference, 2009. University of Zurich and ETH.

16. Turner, T.C., et al., Picturing Usenet: Mapping Computer-Mediated Collective Action. Journal of Computer-Mediated Communication, 2005. 10(4).

17. Klamma, R., et al., Pattern-Based Cross Media Social Network Analysis for Technology Enhanced Learning in Europe. Lecture Notes in Computer Science, 2006. 4227: p. 242-256.

18. Leydesdorff, L. and T. Schank, Dynamic animations of journal maps: Indicators of structural changes and interdisciplinary developments. Journal of the American Society for Information Science and Technology, 2008. 59(11): p. 1810-1818.

19. Burt, R.S., Structural Holes and Good Ideas. The American Journal of Sociology, 2004. 110(2): p. 349-399.

20. Stegbauer, C. and A. Rausch, Moving Structure: Möglichkeiten der positionalen Analyse von Verlaufsdaten am Beispiel von Mailinglisten, in Strukturalistische Internetforschung. Netzwerkanalysen internetbasierter Kommunikationsräume, C. Stegbauer and A. Rausch, Editors. 2006, VS: Wiesbaden. p. 169-198.

21. Kronegger, L., A. Ferligoj, and P. Doreian, On the dynamics of national scientific systems. Quality & Quantity, 2011: p. 1-27.

22. Mote, J.E., R&D ecology: using 2-mode network analysis to explore complexity in R&D environments. Journal of Engineerings and Technology, 2005. 22: p. 93-111.

23. Mote, J.E., et al., New Directions in the Use of Network Analysis in Research and Product Development Evaluation. Research Evaluation, 2007. 16(3): p. 191-203.

24. Rogers, J., B. Bozeman, and I. Chompalov, Obstacles and opportunities in the application of network analysis to the evaluation of R&D. Research Evaluation, 2001. 10(3): p. 161-172.

25. Newman, M.E.J. and M. Girvan, Finding and evaluating community structure in networks. Physical Review E, 2004. 69(2): p. 026113.

26. Waltman, L., N.J. Van Eck, and E.C.M. Noyons, A unified approach to mapping and clustering of bibliometric networks. Journal of Infometrics, 2010. 4(4): p. 629-635.

27. Derényi, I., G. Palla, and T. Vicsek, Clique percolation in random networks. Phys. Rev. Lett., 2005. 94: p. 160202.

28. Wang, Q. and E. Fleury, Mining time-dependent communities. Latin-American Workshop on Dynamic Networks, 2010. Available online: http://hal.archives-ouvertes.fr/docs/00/53/17/35/PDF/lawdn2010_submission_1.pdf.

29. Palla, G., A.-L. Barabasi, and T. Vicsek, Quantifying social group evolution. Nature, 2007. 446(7136): p. 664-667.

Copyright © SISOB Consortium 2011-2013 Page 36/59

Page 44: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

30. Greene, D., D. Doyle, and P. Cunningham. Tracking the evolution of communities in dynamic social networks. in International Conference on Advances in Social Networks Analysis and Mining (ASONAM’10). 2010. Odense, Denmark.

31. Chakrabarti, D., R. Kumar, and A. Tomkins. Evolutionary clustering. in KDD '06: Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining. 2006. Philadelphia, USA: ACM Press.

32. Freeman, L.C., Centrality in social networks conceptual clarification. Social Networks, 1978. 1(3): p. 215-239.

33. Huisman, M. and M.A. van Duijn, Software for Social Network Analysis, in Models and Methods in Social Network Analysis, P.J. Carrington, J. Scott, and S. Wasserman, Editors. 2004, Cambridge University Press: New York. p. 270-316.

34. Freeman, L.C., Graphical techniques for exploring social netowrk data, in Models and methods in social network analysis, P.J. Carrington, J. Scott, and S. Wasserman, Editors. 2004, Cambridge University Press: Cambridge, NY.

35. Trier, M., Towards a Social Network Intelligence Tool for Visual Analysis of Virtual Communication Networks, in Virtuelle Organisationen und Neue Medien, K. Meissner and M. Engelien, Editors. 2006, TUDpress Verlag: Dresden, http://www.commetrix.de/MTrier-Social%20Network%20Intelligence%20Tool%20Virtual%20Communication%20Networks-GENEME2006-PREFINAL_DRAFT.pdf.

36. Groh, G., H. Hanstein, and W. Wörndl. Interactively visualizing dynamic social networks with dyson. 2009. In Proc. of the IUI’09 Workshop on Visual Interfaces to the Social and the Semantic Web.

37. Harrer, A., S. Zeini, and S. Ziebarth, Visualisation of the Dynamics for Longitudinal Analysis of Computer-mediated Social Networks-concept and Exemplary Cases, in From Sociology to Computing in Social Networks, N. Memon and R. Alhajj, Editors. 2010, Springer Vienna. p. 119-134.

38. Laender, A.H.F., et al., A Brief Survey of Web Data Extraction Tools. ACM Sigmod Record, 2002: p. 3184-93.

39. Hansen, D., B. Shneiderman, and M.A. Smith, Analyzing Social Media Networks with NodeXL: Insights from a Connected World, 2010: Morgan Kaufmann.

40. Maimon, O. and L. Rokach, Data Mining and Knowledge Discovery Handbook. 2nd ed, 2010: Springer, ISBN: 978-0-387-09822-7.

41. Larose, D.T., Frontmatter. Data Mining Methods and Models, 2005: John Wiley & Sons, Inc. i-xvi.

42. Grossman, R.L., M.F. Hornick, and G. Meyer, Data mining standards initiatives. Commun. ACM, 2002. 45(8): p. 59-61.

43. Kadav, A., J. Kawale, and P. Mitra, Data Mining Standards. DataMiningGrid, 2003. http://www.datamininggrid.org/wdat/works/att/standard01.content.08439.pdf.

44. OECD, Mobilising human resources for innovation, 2000, Paris: OECD.45. OECD, International mobility of the highly skilled, 2002, Paris: OECD.

Copyright © SISOB Consortium 2011-2013 Page 37/59

Page 45: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

46. European Commission, High-level Expert Group on Improving Mobility of Researchers, 2001, Final report: Brussels.

47. Granovetter, M., Economic action and social structure: The problem of embeddedness. American Journal of Sociology, 1985. 91: p. 481-510.

48. Griliches, Z., Research expenditures and growth accounting, in Science and technology in economic growth, B.R. Williams, Editor, 1973, Macmillan: London. p. 59-95.

49. Schultz, T., Investment in human capital. The american Economic Review, 1961. 51(1): p. 1-17.

50. Schultz, T., The economic value of education, 1963, Columbia: Columbia University Press.

51. Schultz, T., Restoring economic equilibrium: human capital in the modernizing economy, 1990, Oxford: Blackwell.

52. Nelson, R. and E. Phelps, Investment in humans, technological diffusion and economic growth. American Economic Review, 1966. 61: p. 69-75.

53. Becker, G.S., Irrational Behavior and Economic Theory. The Journal of Political Economy, 1962. 70(1): p. 1-13.

54. Becker, G.S., Human Capital1964, New York: Columbia University Press (for NBER).

55. Bourdieu, P., The forms of social capital, in Handbook of theory and research for the sociology of education, J.G. Richardson, Editor 1986, Greenwood: New York.

56. Coleman, J.C., Social capital in the creation of human capital. American Journal of Sociology, 1988. 94: p. 95-120.

57. Burt, R.S., A note on social capital and network content. Social Networks, 1997. 19(4): p. 355-373.

58. Hargens, L.L. and G.M. Farr, An examination of recent hypotheses about institutional inbreeding. American Journal of Sociology, 1973. 78: p. 1381-1402.

59. Cruz-Castro, L. and L. Sanz-Menedez, Mobility versus job stability: Assessing tenure and productivity outcomes. Research Policy, 2010. 39,: p. 27-38.

60. Agrawal, A., I. Cockburn, and J. McHale, Gone but not forgotten: knowledge flows, labor mobility, and enduring social relationships. Journal of Economic Geography, 2006. 6(5): p. 571-591.

61. Almeida, P. and B. Kogut, Localization of Knowledge and the Mobility of Engineers in Regional Networks. Management Science, 1999. 45(7): p. 905-917.

62. Breschi, S. and F. Lissoni, Mobility and social Networks: Localised Knowledge Spillovers Revisited2003: CESPRI Working Paper No. 142.

63. Azoulay, P., J.S. Graff Zivin, and B.N. Sampat, The Diffusion of Scientific Knowledge Across Time and Space: Evidence from Professional Transitions for the Superstars of Medicine. NBER Working Paper No. 16683, 2011.

64. Hoisl, K., Tracing mobile inventors--The causality between inventor mobility and inventor productivity. Research Policy, 2007. 36(5): p. 619-636.

Copyright © SISOB Consortium 2011-2013 Page 38/59

Page 46: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

65. Zubieta, A.F., Recognition and weak ties: is there a positive effect of postdoctoral position on academic performance and career development? Research Evaluation, 2009. 18: p. 105-115.

66. van Heeringen, A. and P.A. Dijkwel, Effect of mobility on productivity. Scientometrics, 1987. 11: p. 267-280.

67. Allison, P.D. and J.S. Long, Departmental Effects on Scientific Productivity. American Sociological Review, 1990. 55(4): p. 469-478.

68. Kim, E.H., A. Morse, and L. Zingales, Are elite universities losing their competitive edge? Journal of Financial Economics, 2006. 93: p. 353-381.

69. Coupé, T., V. Smeets, and F. Warzynski, Incentives, Sorting and Productivity along the Career: Evidence from a Sample of Top Economists. Journal of Law, Economics, and Organization, 2006. 22(1): p. 137-167.

70. Cañibano, C., J. Otamendi, and I. Andujar, Measuring and assessing researcher mobility from CV analysis: the case of the Ramón y Cajal programme in Spain. Research Evaluation, 2008. 17: p. 17-31.

71. Hoare, A.G., Transferred Skills and University Excellence?: An Exploratory Analysis of the Geography of Mobility of UK Academic Staff. Geografiska Annaler. Series B, Human Geography, 1994. 76(3): p. 143-160.

72. Crane, D., Scientists at Major and Minor Universities: A Study of Productivity and Recognition. American Sociological Review, 1965. 30(5): p. 699-714.

73. Crane, D., The Academic Marketplace Revisited: A Study of Faculty Mobility Using the Cartter Ratings. The American Journal of Sociology, 1970. 75(6): p. 953-964.

74. Long, J.S., Productivity and Academic Position in the Scientific Career. American Sociological Review, 1978. 43(6): p. 889-908.

75. Baldi, S., Prestige Determinants of First Academic Job for New Sociology Ph.D.s 1985-1992. The Sociological Quarterly, 1995. 36(4): p. 777-789.

76. Debackere, K. and M.A. Rappa, Scientists at major and minor universities: mobility along the prestige continuum. Research Policy, 1995. 24(1): p. 137-150.

77. Chan, K.C., C.R. Chen, and T.L. Steiner, Production in the Finance Literature, Institutional Reputation, and Labor Mobility in Academia: A Global Perspective. Financial Management, 2002. 31(4): p. 131-156.

78. Burris, V., The Academic Caste System: Prestige Hierarchies in PhD Exchange Networks. American Sociological Review, 2004. 69(2): p. 239-264.

79. Dietz, J.S. and B. Bozeman, Academic careers, patents, and productivity: industry experience as scientific and technical human capital. Research Policy, 2005. 34(3): p. 349-367.

80. Zucker, L.G., M.R. Darby, and M. Torero, Labor mobility from academe to commerce. Journal of Labor Economics, 2002. 20: p. 629-660.

81. Crespi, G., A. Geuna, and L. Nesta, The mobility of university inventors in Europe. The Journal of Technology Transfer, 2007. 32(3): p. 195-215.

Copyright © SISOB Consortium 2011-2013 Page 39/59

Page 47: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

82. Pezzoni, M., V. Sterzi, and F. Lissoni, Career progress in centralized academic systems: an analysis of French and Italian physicists. KITeS Working Paper No. 26, 2009.

83. Zinovyeva, N., Academic Inbreeding in Spain: Evidence from a Randomized Natural Experiment. Mimeo, 2011.

84. Marcson, S., The scientist in American industry, 1960, Princeton: Princeton Unversity Press.

85. Krohn, R.G., The Institutional Location of the Scientist and His Scientific values. Engineering Management, IRE Transactions on, 1961. EM-8(3): p. 133-138.

86. Kornhausser, W., Scientists in industry. Conflict and accommodation1962, Berkeley: University of California.

87. Hagstrom, W.O., The scientific community, 1965, New York: Basic Books.88. Stephan, P., Job Market Effects on Scientific Productivity. Mimeo, 2006.89. Melin, G., The dark side of mobility: negative experiences of doing a postdoc

period abroad. Research Evaluation, 2005. 14: p. 229-237.90. Valente, T.W., Network models and methods for studying the diffusion of

innovations, in Models and methods in social network analysis, P.J. Carrington, J. Scott, and S. Wasserman, Editors. 2005, Cambridge University Press: New York, NY.

91. Rogers, E.M., Diffusion of innovations,1962, New York, NY: Free Press.92. Gibbons, M., et al., The new production of knowledge: the dynamics of science

and research in contemporary societies, 1994, London: Sage.93. Star, S.L. and J.R. Griesemer, Institutional Ecology, 'Translations' and Boundary

Objects: Amateurs and Professionals in Berkeley's Museum of Vertebrate Zoology, 1907-39. Social Studies of Science, 1989. 19(3): p. 387-420.

94. Schummer, J., Multidisciplinarity, interdisciplinarity, and patterns of research collaboration in nanoscience and nanotechnology. Scientometrics, 2004. 59(3): p. 425-465.

95. Heimeriks, G., M. Hörlesberger, and P. Van Den Besselaar, Mapping communication and collaboration in heterogeneous research networks. Scientometrics, 2003. 58(2): p. 391-413.

96. Nonaka, I. and H. Takeuchi, The knowledge creating company: how Japanese companies create the dynamics of innovation, 1995, New York: Oxford University Press.

97. Wagner, C.S. and L. Leydesdorff, Network structure, self-organization, and the growth of international collaboration in science. Research Policy, 2005. 34(10): p. 1608-1618.

98. Albert, R. and A. Barabási, Statistical mechanics of complex networks. Reviews of Modern Physics, 2002. 74(1): p. 47.

99. Braun, T., W. Glänzel, and A. Schubert, Publication and cooperation patterns of the authors of neuroscience journals. Scientometrics, 2001. 50(3): p. 499-510.

100. El Louadi, M., Knowledge heterogeneity and social network analysis – Towards conceptual and measurement clarifications. Knowledge Management Research & Practice, 2008. 6: p. 199-213.

Copyright © SISOB Consortium 2011-2013 Page 40/59

Page 48: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

101. Jansen, D., R. von Görtz, and R. Heidler, Knowledge production and the structure of collaboration networks in two scientific fields. Scientometrics, 2010. 83(1): p. 219-241.

102. Leydesdorff, L. and I. Rafols, A global map of science based on the ISI subject categories. Journal of the American Society for Information Science and Technology, 2009. 60(2): p. 348-362.

103. Van Den Besselaar, P. and G. Heimeriks, Mapping research topics using word-reference co-occurrences: a method and an exploratory case study. Scientometrics, 2006. 68(3): p. 377-393.

104. Bullinger, A.C., et al. Towards research collaboration – a taxonomy of social research network sites. in Sixteenth Americas Conference on Information Systems (AMCIS). 2010. Lima.

105. Campanario, J.M., The journal scout. The Scientist, 1997. 11(9).106. Armstong, J.S., Management Science: What Does It Have to do With Management

or Science? Marketing Bulletin, 1998. 9(1): p. 1-13.107. Spier, R., The history of the peer-review process. Trends in Biotechnology, 2002.

20(8): p. 357-358.108. Armstrong, J.S., Peer Review for Journals: Evidence on Quality Control, Fairness,

and Innovation. Science and Engineering Ethics, 1997. 3: p. 63-84.109. Campanario, J.M., Peer review for journals as it stands today-Part 1. Science

Communication, 1998. 19: p. 181-211.110. Benos, D.J., et al., The ups and downs of peer review. Advances in Physiology

Education, 2007. 31(2): p. 145-152.111. editorial, A., Coping with peer rejection. Nature, 2003. 425(6959): p. 645.112. Nature Pros and cons of open peer review. Nature Neuroscience, 1999. 2, 197-

198.113. Bradley, J.V., Pernicious publication practices. Bulletin of the Psychonomic

Society, 1981. 18: p. 31-34.114. Fletcher, R.H. and S.W. Fletcher, Evidence for the effectiveness of peer review. Sci

Eng Ethics, 1997. 3: p. 35-43.115. MacNealy, M.S., B.W. Speck, and N. Clements, Publishing in technical

communication journals from the successful author's point of view. Technical communication, 1994. 41(2): p. 240-259.

116. Relman, A.S., Are we a filter or a sponge? New England Journal of Medicine, 1978. 299: p. 197.

117. Scanlan, B.P., Passengers, drivers and traffic cops on peer review road. Serials Librarian, 1991. 19: p. 71-73.

118. Yoels, W.C., Doctoral origins and appointment pattems of editors of the American Sociological Review, 1948-1968. Amercian Sociologist 1971. 6: p. 134-139.

119. Travis, G.D.L. and H.M. Collins, New Light on Old Boys: Cognitive and Institutional Particularism in the Peer Review System. Science, Technology, & Human Values, 1991. 16(3): p. 322-341.

Copyright © SISOB Consortium 2011-2013 Page 41/59

Page 49: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

120. Hamermesh, D.S., Facts and Myths about Refereeing. The Journal of Economic Perspectives, 1994. 8(1): p. 153-163.

121. Mahoney, M.J., Publication prejudices: An experimental study of confirmatory bias in the peer review system. Cognitive Therapy and Research, 1977. 1(2): p. 161-175.

122. Popper, K.R., Popper: Conjectures and Refutations, 1963, London: Routledge & Kegan Paul.

123. Kerr, S., J. Tolliver, and D. Petree, Manuscript Characteristics Which Influence Acceptance for Management and Social Science Journals. The Academy of Management Journal, 1977. 20(1): p. 132-141.

124. Beyer, J.M., G.C. Roland, and W.B. Fox, The Review Process and the Fates of Manuscripts Submitted to AMJ. The Academy of Management Journal, 1995. 38(5): p. 1219-1260.

125. Guide for Applicants: Collaborative projects, 2009, Cordis, http://cordis.europa.eu.

126. Atkinson, D.R., M.J. Furlong, and B.E. Wampold, Statistical significance, reviewer evaluations , and the scientific process: Is there a statistically significant relationship? Journal of Counseling Psychology, 1982. 29(2): p. 189-194.

127. Armstrong, J.S., Unintelligible management research and academic prestige. Interfaces, 1980. 10(April): p. 80-86.

128. Lloyd, M.E., Gender factors in reviewer recommendations for manuscript publication. Journal of Applied Behavior Analysis, 1990. 23: p. 539-543.

129. Wenneras, C. and A. Wold, Nepotism and Sexism in Peer-Review. Nature, 1997. CCCLXXXVII: p. 341-343.

130. Sandström, U. and M. Hällsten, Persistent nepotism in peer-review. Scientometrics, 2008. 74(2): p. 175-189.

131. Link, A.M., US and Non-US Submissions. JAMA: The Journal of the American Medical Association, 1998. 280(3): p. 246-247.

132. Peters, D.P. and S.J. Ceci, Peer-review practices of psychology journals: The fate of published articles, submitted again. The Behavioral and brain Sciences, 1982. 5: p. 187-195.

133. Wade, N., Peer review systems: How to hand our money fairly. Science, 1973. 179: p. 158-61.

134. Lerner, E.J., Fraud shows peer review flaws. Industrial Physicist, 2003: p. 12-17.135. Nature, Can peer review police fraud? Nature Neuroscience, 2006. 9(2): p. 149.136. Cole, S., J.R. Cole, and G.A. Simon, Chance and Consensus in Peer Review. Science,

1981. 214(4523): p. 881-886.137. Lehmann, S., A.D. Jackson, and B.E. Lautrup, Measures for measures. Nature,

2006. 444(7122): p. 1003-1004.138. Frontiers. Review Guidelines,

http://www.frontiersinneuroscience.org/analytical_and_experimental_pharmacology/reviewguidelines. 2011 [cited 2011 April 12].

139. Groves, T., Best practice in peer review and editing, ensuring article quality. Notfall + Rettungsmedizin, 2010. 13(1): p. 6-8.

Copyright © SISOB Consortium 2011-2013 Page 42/59

Page 50: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

140. Sivakumar, B., Peer-Review System and Anonymity of Reviewers: A Three-Pronged Proposal. J. Hydrologic Engrg., 2008. 13(7): p. 529-530.

141. Goldbeck-Wood, S., Evidence on peer review—scientific quality control or smokescreen? BMJ, 1999. 318(7175): p. 44-45.

142. Nature., Editorial Practices: A Guide to Publication Policies of the Nature Journals. Nature, 2009. http://www.nature.com/authors/gta.pdf.

Copyright © SISOB Consortium 2011-2013 Page 43/59

Page 51: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

Appendix A: Common Network Indicators

IntroductionThe aims of this appendix are (1) to list the most common measures and indicators used in complex network analysis at the technical level (2) to impose a taxonomy on these concepts and (3) to prepare an overview for a subsequent Deliverable, concerning the actual meaning and use of these indicators in research on science and technology.

The appendix is organized in the following sections:

Basic network indicators. This section contains a structured list of the most common network indicators in the literature.

Combined indicators. This section proposes taxonomy of measures constructed by combining basic measures. Here we describe the typical ways in which basic network measures are combined in the literature. The interpretation and use of these tools in STS and policy studies will be discussed in a future. Deliverable.

In what follows we use the following abbreviations:

(W) Weighted graphs

(D) Directed graphs

Basic network indicators

Indicators for the network (subnetwork) level

Distributional indicatorsStatistics (distributions) of structural node, edge (link), properties and communities at the network level

Degree distributionDistribution of individual node degree values (cf. degree centrality)

CentralizationDistribution of individual node centrality values (cf. various centrality indices)

Edge weight distribution (W) Distribution of edge weight valuesSize distribution of connected components

Cf. connected components

Size distribution of Detected Cf. community structure

Copyright © SISOB Consortium 2011-2013 Page 44/59

Page 52: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

communities

Indicators of overall structure

Single-valued indicators of overall structure

DensityRatio of the number of potential and actual ties in a graph

DiameterThe length of the maximum path in a graph, whereby a path between two nodes a and b is a series of ties jointly connecting a and b.

Average path length The mean value of path lengths for a graph.Global and average clustering coefficient / Transitivity

Network-level measures of the degree to which nodes in a graph form fully connected groups (cf. density)

ConnectednessFor a graph, the degree of containing mutually isolated subgraphs (connected components).

Dyadic/Triadic censusNumber of dyads/triads in a graph as that of two- and three-element subgraphs, respectively.

Similarity(between networks)

Similarity measures for network structure in terms of matrix similarity.

Indicators for the node/edge level

Positional (centrality) measures

Degree centrality For any node, the number of edges connecting with the node/the number of direct neighbors

Indegree centrality (D) For any node, the number of incoming edges/linksOutdegree centrality (D) For any node, the number of outgoing edges/links

Betweenness centrality For any node, the number of shortest paths the node is crossed by.

Closeness centrality For any node, the mean length of the shortest paths from the node to all other nodes reachable from it.

Bonacich/Eigenvector centrality

For any node, the sum of recursively assigned scores of its ties, the latter calculated according to the principle that ties to well-connected nodes score higher than ties to nodes with fewer connections.

Path length measures

Copyright © SISOB Consortium 2011-2013 Page 45/59

Page 53: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

Shortest path between nodes

(cf. diameter)

Node Attributes

Non-relational properties (variables) assigned to individual nodesEdge Attributes

Non-relational properties (variables) assigned to individual edges

Indicators of community structure

Subgraph statistics

Number, size distribution of

Connected components connected sets of nodes isolated from other such sets

Cliques

(k-cliques, k-communities etc.)

Fully connected subgraphs with k nodes

NeighborhoodsThe set of nodes a particular element of the network is directly connected to (with path length=1)

Detected partitionsMutually exclusive sets of nodes, identified by some community detection method each

Detected overlapping communities

Overlapping sets of nodes identified by some community detection method each

Structural equivalence statistics

Number of equivalence classes: equivalence classes are sets of nodes with identical structural features (being connected to the same nodes, having identical centrality values etc.)

Combined indicatorsCombined indicators in the literature most often relate node attributes to structural node properties. Major types are:

Copyright © SISOB Consortium 2011-2013 Page 46/59

Page 54: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

Correlates of positional features

Measures of the associations between positional node properties (e.g. betweenness centrality of a node) and node attributes (some non-relational variable value of the same node)

Tie or attachment types and diversity

Measures of the relations between node properties of structurally related nodes (neighbors, nodes related by different types of paths etc.)

Network diversity measures

Diversity exhibited by networks measured as some function of (1) node attributes [e.g. size] and (2) some indicator of node relatedness (e.g. distance).

Community profiles

Distributional properties or characterization of all types of communities measured in terms node values

Copyright © SISOB Consortium 2011-2013 Page 47/59

Page 55: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

Appendix B: Software tools for KDD

Open Source Febrl (Freely Extensible Biomedical Record Linkage) [Christen-2009]

o https://sourceforge.net/projects/febrl/ o Preprocessing. It includes a variety of functionalities required for data

cleaning, deduplication and record linkage, and it provides a graphical user interface that facilitates its application for users who do not have programming experience.

R: Statisticso http://www.r-project.org o R-EXTENSIONS

http://www.r-project.org/other-projects.html o R-WEKA. [Hornik-2009]

http://cran.r-project.org/web/packages/RWeka/index.html To make the different sets of tools from both environments

available in a single unified system, an R package RWeka is suggested which interfaces Weka's functionality to R.

Augustus [Chaves-2006]o http://opendatagroup.com/services/open-source/augustus-scoring-

engineo PMML 4-compliant scoring engine that works with segmented models.

Open Data developed and continues to maintain a suite of analytic tools. Designed for use with statistical and data mining models. Written in python. Augustus is part of a comprehensive suite of tools focused on change detection.

KNIME (Konstanz Information Miner)[Berthold-2009]o http://www.knime.org o It is a modular environment, which enables easy visual assembly and

interactive execution of a data pipeline. The user can model workflows, which consist of nodes that process data, transported via connections between those nodes

o KNIME-EXTENSIONS http://labs.knime.org

RapidMiner [Mierswa-2006][RAPIDMINER-2009]o http://rapid-i.com/content/view/181/190 o It is an environment for machine learning and data mining processes. A

modular operator concept allows the design of complex nested operator chains for a huge number of learning problems. The data handling is transparent to the operators.

Rattle [Williams-2009]

Copyright © SISOB Consortium 2011-2013 Page 48/59

Page 56: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

o http://rattle.togaware.com o A Data Mining GUI for R. The Rattle package provides a graphical user

interface specifically for data mining using R. It also provides a stepping stone toward using R as a programming language for data analysis

WEKA [Hall-2009][Bouckaert-2010]o http://www.cs.waikato.ac.nz/ml/weka/index.html o The WEKA project aims to provide a comprehensive collection of machine

learning algorithms and data preprocessing tools to researchers and practitioners alike. It allows users to quickly try out and compare different machine learning methods on new data sets. Its modular, extensible architecture allows sophisticated data mining processes to be built up from the wide collection of base learning algorithms and tools provided. Extending the toolkit is easy thanks to a simple API, plugin mechanisms and facilities that automate the integration of new learning algorithms with WEKA's graphical user interfaces Tool for Data Mining

o Comparison R vs. WEKA [PENTAHO-2007]o WEKA-EXTENSIONS

http://weka.wikispaces.com/Related+Projects o MOA [Bifet-2010]

http://moa.cs.waikato.ac.nz Data Stream Mining

Pentaho (based on weka)o http://www.pentaho.com o Business Intelligence Suite (Reporting, Analysis, Dashboards, Data

Integration (ETL), Data Mining) XELOPES [PRUSYS-2009]

o http://www.prudsys.de/en/technology/xelopes o Business Intelligence library. Implementations for Java, C++ and C#,

CORBA and web service interfaces are currently available Orange [Demsar-2004]

o http://orange.biolab.si o Open source data visualization and analysis. Data mining through visual

programming or Python scripting METAL [Berrer-2000]

o http://www.metal-kdd.org o A Meta-Learning Assistant for Providing User Support in Machine Learning

and Data. This research is supported by the European Union (ESPRIT Reactive LTR 26.357).

Shogun [Sonnenburg-2010]o http://www.shogun-toolbox.org o Machine Learning Toolbox. Implemented in C++ and interfaces to MATLAB,

R, Octave, Python PyBrain [Schaul-2010]

Copyright © SISOB Consortium 2011-2013 Page 49/59

Page 57: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

o http://www.pybrain.org o Machine learning library for Python

Dlib-ml [King-2009]o [http://dlib.net/ml.html|http://dlib.net/ml.html]]o Machine learning library for C++

Java-ML [Aabel-2009]o [http://java-ml.sourceforge.net|http://java-ml.sourceforge.net]]o Machine learning library for Java

JHepWork [Chekanov-2010]o http://jwork.org/jhepwork o Machine learning library for Jython.

LibDAI [Mooij-2010]o http://people.kyb.tuebingen.mpg.de/jorism/libDAI o Library for discrete approximate inference in graphical models. C++

BIRT (Business Intelligence and Reporting Tools)o http://www.eclipse.org/birt o Reporting and business intelligence capabilities for rich client and web

applications, especially those based on Java and Java EE (ECLIPSE) JMLR MLOSS (Machine Learning Open Source Software) from JMLR

o http://jmlr.csail.mit.edu/mloss o Implementations of non-trivial machine learning algorithms, toolboxes, etc.

Commercial (DM, KDD and BI – business intelligence --)In the chapter 65 of [Maimon-2011] it is a collection of modern SW. We have updated and completed that collection. Most of the tools are placed in the context of Business Intelligence and big companies in the field of DBMS usually incorporate some extension or plugin to support data mining tasks.

Microsoft SQL Servero http://msdn.microsoft.com/en-us/library/ms146608.aspx o BOOK: Smart Business Intelligence Solutions with Microsoft SQL Server

2008 (Safari)o BOOK: Microsoft SQL Server 2008 Bible (Safari)

IBM SPSS Modelero http://www-01.ibm.com/software/analytics/spss/products/modeler o http://www.spss.com/software/modeler

SAS Enterprise Miner [SAS-2003]o http://www.sas.com/technologies/analytics/datamining/miner/index.html

Oracle Data Miner [ORACLE-2011]o http://www.oracle.com/technetwork/database/options/odm/index.html

KXENo http://www.kxen.com

Copyright © SISOB Consortium 2011-2013 Page 50/59

Page 58: sisob.lcc.uma.essisob.lcc.uma.es/repositorio/deliverables/SISOB D2.1 r… · Web viewsisob.lcc.uma.es

http://sisob.lcc.uma.es

SPM (Salford Predictive Modeling Suite)o http://salford-systems.com

SAND InDatabase-Analyticso http://www.sand.com/analytics/indatabase-analytics

Teradata Warehouse Minero http://www.teradata.com/t/products-and-services/teradata-warehouse-

miner Predixion Insight

o https://www.predixionsoftware.com/predixion/Products.aspx Megaputer PolyAnalyst

o http://www.megaputer.com/data_mining.php Inforsense

o http://www.inforsense.com MicroStrategy

o http://www.microstrategy.com TIBCO Spotfire Miner

o http://spotfire.tibco.com/products/data-mining-applications.aspx Pervasive DataRush (Apache Hadoop cluster)

o http://www.pervasivedatarush.com/home.aspx Zementis ADAPA [Guazelli-2009a]

o http://www.zementis.com o Scoring engine in the CLOUD.o Software as a Service on the Amazon Cloud (NOT FREE)o ADAPA (Adaptive Decision And Predictive Analytics) Add-in for Microsoft

Office Excel

Other software relevant to SISOB (Web Reviews)

Software Suites for Data Mining, Analytics, and Knowledge Discoveryo http://www.kdnuggets.com/software/suites.html o http://www.kdnuggets.com/software/classification.html o http://www.kdnuggets.com/software/clustering.html o http://www.kdnuggets.com/software/data-cleaning.html

Social network analysis, Link Analysis, and Visualizationo http://www.kdnuggets.com/software/social-network-analysis.html

Visualization and Data Mining Softwareo http://www.kdnuggets.com/software/visualization.html

Network text analysis, Network dynamics and visualization:o http://www.textrend.org/

Copyright © SISOB Consortium 2011-2013 Page 51/59