Skip to main content
PLOS One logoLink to PLOS One
. 2022 Nov 9;17(11):e0276061. doi: 10.1371/journal.pone.0276061

In search of diverse and connected teams: A computational approach to assemble diverse teams based on members’ social networks

Diego Gómez-Zará 1,2,3,*, Archan Das 4, Bradley Pawlow 5, Noshir Contractor 5,6,7,8
Editor: Seyedali Mirjalili9
PMCID: PMC9645621  PMID: 36350821

Abstract

Previous research shows that teams with diverse backgrounds and skills can outperform homogeneous teams. However, people often prefer to work with others who are similar and familiar to them and fail to assemble teams with high diversity levels. We study the team formation problem by considering a pool of individuals with different skills and characteristics, and a social network that captures the familiarity among these individuals. The goal is to assign all individuals to diverse teams based on their social connections, thereby allowing them to preserve a level of familiarity. We formulate this team formation problem as a multi-objective optimization problem to split members into well-connected and diverse teams within a social network. We implement this problem employing the Non-dominated Sorting Genetic Algorithm II (NSGA-II), which finds team combinations with high familiarity and diversity levels in O(n2) time. We tested this algorithm on three empirically collected team formation datasets and against three benchmark algorithms. The experimental results confirm that the proposed algorithm successfully formed teams that have both diversity in member attributes and previous connections between members. We discuss the benefits of using computational approaches to augment team formation and composition.

Introduction

Forming teams today is different from past decades. Nowadays, organizations and institutions aim to assemble groups based not only on members’ expertise but also on diversity criteria [1, 2]. Because the workforce is becoming increasingly more diverse, more organizations are committing to bringing together members from different educational backgrounds, functional backgrounds, and demographic attributes in the same team [3, 4].

Numerous studies show the potential benefits of diversity in teams [5, 6]. At the identity level, research shows that demographic diversity—team members of different gender, culture, race, etc.—can boost team performance. Cultivating demographic diversity in teams can bring different traits, points of view, and experiences inherent to the demographic group [7, 8]. Some examples are gender diversity, which promotes productivity in software development teams [9], teams’ collective intelligence [10], and innovations in R&D groups [11]. One study showed that racial diversity can also bring alternative perspectives and stimulate creativity, generating more original and competitive ideas [12]. Cultural diversity is another example: it helps teams produce more creative results than culturally homogeneous teams [13]. At the cognitive level, teams with high levels of functional diversity—that is, team members with different expertise, careers, and backgrounds—can deliver more original and creative outcomes. Promoting functional diversity can enhance creativity because it expands the breadth of information, knowledge, ideas, and perspectives within a team [14]. It also encourages divergent thinking, greater scope of skills, and idea recombination [15, 16]. As a result, functionally diverse teams are more likely to solve complex problems that require creativity and innovation than homogeneous groups [2, 8, 16]. Overall, the interplay of demographic and functional diversity plays a role in how team members’ differences leverage their work and performance [7].

Despite the potential benefits of diversity in teams, research also shows that diversity is a “double-edged sword” [17]. Prior studies offer mixed, and even contradictory, results of the effects of diversity on teams [14, 17, 18]. While functional diversity can cause coordination problems and conflicts in a group due to differences in training and knowledge, demographic diversity can elicit inter-bias among members (i.e., “us-them” distinction) [19], leading to a lack of cohesion, communication, and trust [2022]. For decades, organizations have promoted diversity training to help members work with others who are different from them. Although, when people are assigned to work in a diverse team, they are less likely to engage with the team and be motivated to work with teammates that differ in demographic or functional attributes [23].

One potential solution to moderate the adverse effects of diversity on teams is enabling team familiarity (i.e., team members’ prior experience working with one another). A substantial body of literature shows that prior collaboration leads to a greater likelihood of success and future collaborations [2426]. Team familiarity creates the foundations of trust, information distribution, and communication among members [27, 28]. And because team familiarity aids members in locating, sharing, and distributing their knowledge, team familiarity may address many problems created by diversity without compromising its potential benefits [29].

Can organizations assemble teams with high diversity levels and familiarity simultaneously? Rather than forming teams based on either diversity criteria or prior relationships, combining both can help members promote trust, and organizations make the benefits of diversity more salient [29]. In this work, we propose a computational approach to discover suitable team combinations that maximize team diversity and familiarity at the same time. We chose these two team characteristics because both can be determined during the team formation process. Since this task requires assessing all the possible combinations among the available members, we elaborate on an optimization problem and its algorithm implementations to find invaluable team combinations efficiently.

We formulate this team formation problem as a multi-objective optimization problem to assemble teams maximizing their diversity and familiarity simultaneously. We use Harrison and Klein’s framework [30] to calculate teams’ diversity based on the variety and disparity of attributes, and we use Kargar and An’s communication cost metric [31] to calculate teams’ familiarity based on members’ social network structure [32]. We then implement this problem employing the Non-dominated Sorting Genetic Algorithm II (NSGA-II). This implementation is appropriate because it provides a set of efficient team combinations and considers the trade-offs of different objectives. We demonstrate the effectiveness of our approach using three datasets that contain team membership information: (1) students self-assembling teams using the MyDreamTeam platform [33], (2) scientists co-authoring papers provided by the bibsonomy dataset [34], and (3) teams collaborating on GitHub provided by the GHTorrent dataset [35]. We assess our proposed algorithm against other multi-objective optimization methods highly cited in the literature by evaluating its solutions and running time. The results demonstrate that our proposed algorithm successfully provided solutions with higher diversity and familiarity levels.

The main contribution of this paper is the formulation of the team formation problem considering teams’ diversity levels and members’ familiarity simultaneously. While most studies in team formation algorithms have considered members’ skills or personal costs as team formation objective functions [36], we formulate this optimization problem based on different operationalizations of diversity (i.e., disparity and variety of attributes). The second contribution of this work is the design of algorithms for this team formation problem that assigns all available individuals to a team. Previous team formation problems have mainly focused on finding the best team from a pool and dismissed the remaining individuals [36, 37]. This work also provides theoretical implications for team research. In particular, the use of computational mechanisms to support the team formation processes [3840]. Practical implications of this study contribute to several communities invested in increasing team diversity. Since team builders cannot solve this problem quickly by manually checking each team combination, algorithms can automatize this task by bringing together members who possess existing social connections while, at the same time, from different backgrounds, characteristics, and expertise levels [41, 42]. Expanding the use of this algorithm to broader audiences could provide new benefits for groups that seek to embrace diversity and keep high familiarity levels.

This article is an extended and revised version of a preliminary conference proceeding presented in Complex Networks 2020 [43]. Compared with the conference article, this version (a) presents a review of team formation algorithms, (b) extends the definitions and pseudo-codes of the proposed team formation problem and algorithm, (c) upgrades the proposed algorithm to handle isolated individuals and when the number of available individuals is not a multiple of the team size, (d) evaluates the algorithm with three datasets to prove that our optimization problem can work in other team formation domains, (e) compares its performance against other benchmark multi-objective algorithms, (f) uses quantitative metrics to compare the algorithms’ results, (g) elaborates on the findings and implications of this work for researchers and practitioners, and (h) provides the scripts to pre-process the datasets, the pre-processed datasets, and the scripts with our proposed algorithm and benchmark algorithms for reproducibility purposes.

Related work

Computer science scholars have elaborated different approaches to solving the problem of team assignment [36, 44, 45]. Scholarship has concluded that finding the most efficient team combinations from a pool of individuals is a challenging computational problem, and it is even harder for individuals to solve manually. It is a complex task that requires assessing all the possible combinations among the members of a pool, which can become an insurmountable combinatorial challenge. Given a pool of n members that must be assigned into teams of size k, we must calculate an iterative permutation where we can select the first k members from n, then other k members from nk, and so on. Assuming that k is a multiple of n, we have to calculate n/k permutations. As a result, we must compute n!/(k!n/k*(n/k)!) possible team combinations for a pool of n members. If we want to assemble teams of size three from a class of 18 students, there are 190,590,400 possible combinations (18!/(3!6*6!)). Assessing all these combinations scales in factorial time (i.e., O(n!)). Therefore, this task cannot be done in polynomial time and demands different approaches to find solutions efficiently.

Recent literature reviews [36, 37, 46] characterize team formation algorithms according to three main dimensions: (i) the number of teams that result from the algorithm, (ii) the members’ attributes considered by the algorithm, and (iii) the number of objective functions considered by the algorithm.

Number of teams

Most solutions propose to find the best team possible from a given pool of individuals. The “best-team” approach usually considers the team formation problem as an assignment problem, where the goal is to find the best members who can assemble a team. Key contributions to this literature rely on methodology innovations. For example, El-Ashmawi et al. [47] searches for the team with the least communication costs among team members using an implementation of the particle swarm optimization algorithm. Bhowmik et al. [48] developed a team formation algorithm using a submodular function optimization. This implementation finds the best team of experts with relaxed constraints: teams “must” have some skills while they “should” have others. Lastly, Keane et al. [49] employs a team formation algorithm using a gradient boosting framework to find the minimal team with experts who can work effectively together. A limitation of these methods is that they provide only a single “best” team rather than multiple teams that include all the members from the available pool.

A few studies have explored the problem of assigning all available individuals to teams. One approach is forming multiple teams through iterative heuristics. In this case, teams are assembled by extracting k members of the pool according to an objective function until no more members are left without a team. One example is Agrawal et al. [50], which proposed heuristic algorithms to maximize the gain (or minimize the cost) aggregated over all the teams assembled from a pool of available individuals. This paper presents two iterative heuristic algorithms that team up “strong” members with others who are “weaker” than them. As a result, experts are distributed among several teams. A second approach is formulating the team formation problem as a partition problem. A pool of individuals is partitioned into teams using heuristic metrics for all the assembled teams. Some implementations use clustering algorithms that aim to find members sharing similar characteristics. Some examples are Nurjanah et al.’s implementation [51] that uses Fuzzy C-Means to cluster individuals in homogeneous teams [51], and Srba and Bielikova’s implementation [52] that clusters students according to specific collaborative characteristics. A third approach is finding efficient team combinations using evolutionary algorithms [53, 54]. In a nutshell, evolutionary algorithms start assigning all members to random teams and then alter individuals’ memberships iteratively to find better team combinations. After evaluating the combinations using determined objective functions, evolutionary algorithms keep the best team combinations to find new combinations in the next iteration. One example is Agustín-Blas et al. [53], who developed a genetic algorithm that organizes individuals into groups and searches for team combinations that maximize groups’ required resources.

Members’ attributes

The second dimension focuses on the members’ attributes considered by the algorithm. The algorithms’ goal is to find members that maximize specific team attributes, such as the number of social connections among members or the number of skills covered by the team. Most algorithms set the presence of skills (or expertise) in a team as the main goal. An illustration of this approach is Zakarian and Kusiak’s algorithm [55], which uses mathematical programming to find members that contribute a particular skill to the team. Other computational implementations propose to assign members according to their roles. An algorithm based on this approach is Yannibelli et al. [56], who developed an evolutionary approach that assembles student teams by assigning members to specific roles. Moreover, algorithms can include members’ social networks when forming teams. Lappas et al. [57] searches for the best team with the lowest social distance among members (i.e., a surrogate for communication costs). This article proposes two algorithms called “Rarest First” and “Enhanced Steiner.” Given the problem of assembling the best team of size k with members who have the skills to solve a task T from a social network G, the first algorithm searches for the smallest graph diameter possible among k members who have the skills to solve T. The second algorithm searches for the smallest subset of edges that connects k members with the skills to solve T (i.e., minimum spanning tree). Both algorithms will aim to find the best team possible given a particular social network. Other implementations consider the personnel cost, members’ availability, and workload balance among members as part of the team formation problem [58].

Number of objective functions

The third dimension is the number of objectives being optimized by the team formation algorithm. Some examples are minimizing teams’ communication costs, minimizing teams’ personnel costs, and maximizing the number of skills present in each team. Most algorithms define the team formation problem with a single objective with restrictions [59]. The examples mentioned before follow this single-objective function design. A pitfall is that other beneficial goals for team composition cannot be considered during the optimization process simultaneously (e.g., minimizing communication costs while maximizing the team’s skills).

Prior studies have introduced more than one objective function to the team formation problem. One example is Kargar et al. [60], which presents the “Minimal Cost Contribution” algorithm (MCC). Its goal is to search for the team with the lowest communication costs and the lowest personnel costs simultaneously. MMC’s objective function is a linear combination of both cost functions with a parameter λ that indicates the trade-off between communication and personal costs. This algorithm implements a heuristic approach that adds new members to the team incrementally and considers the costs of adding a new member with respect to the current costs of the assembled team. Despite the benefits of these linear combination formulations, this approach presents two limitations: it provides only one single team solution, and its trade-off variable for the cost functions must be set in advance. Thus, finding other suitable solutions using these methods depends on the adjustment of the trade-off variable, which can add bias to the search process [61].

Recent algorithmic contributions have formulated the team formation problem as a multi-objective optimization problem to optimize two or more objective functions simultaneously [62, 63]. These problems involve trade-offs between two or more objectives since improving a solution in one objective is possible only by making a concession to another objective. Thus, multi-objective optimization problems do not provide a single solution but obtain multiple solutions considering different relevance emphases for the multiple objectives. While in single-objective optimization problems the superiority of one solution over others is determined by the objective function, in multi-objective optimization problems it is determined by dominance. The optimization process looks for solutions that are better than others in all the objective functions. As a result, the problem delivers a set of “non-dominated” solutions, which consists of solutions that can be improved without a simultaneous detriment to at least one of the other objectives. Multi-objective optimization is also known as Pareto optimization. Fig 1 shows an example of a Pareto front showing different non-dominated solutions ranging between two objectives. Computing this Pareto front allows decision makers to compare and check different trade-offs among both dimensions.

Fig 1. Pareto front example.

Fig 1

Non-dominated solutions represent optimal solutions considering trade-offs between the two objective functions.

Based on this approach, multi-objective algorithmic implementations provide a set of team solutions that consider varied assessments of the objective functions [54, 64]. Zhang and Zhang’s implementation [64] selects the members with the highest capabilities for the task and the best interpersonal relationships to assemble the best team. This study uses the particle swarm optimization implementation to determine whether a member i must be part of the best team. Solutions move in a two-dimensional continuous space, and the algorithm applies a sigmoid function to binarize members’ presence. Perez-Toledano et al. [63] developed a genetic algorithm to find competitive basketball teams considering the cost and valuation of each player simultaneously. Each solution consists of a team from a set of available players, and its final Pareto front displays different teams that consider the trade-off between players’ valuation and cost. Based on these formulations, team builders can see and compare other teams and choose what objective they will prioritize when selecting a team.

Problem formulation

After reviewing relevant team formation problems and their respective algorithms, we aim to implement this particular problem that maximizes teams’ diversity and teams’ familiarity simultaneously. This problem is appropriate for a multi-objective optimization formulation since maximizing teams’ familiarity could lead to forming groups with members that are similar to each other [65]. Although we could implement this problem as a single-objective optimization problem, we would have to prioritize one of those goals and avoid trade-offs among solutions. Moreover, prior formulations of the team formation searched for either the best team among multiple objectives or team combinations based on a single objective. We propose a multi-objective optimization problem that assigns all available individuals into teams, resulting in several team combinations that consider different relevance emphases for diversity and familiarity. This work is not the case for previous studies on team formation and provides a new approach to the team formation literature.

Materials and methods

In this section, we introduce the multi-objective problem and definitions that we will use throughout this paper. Our notation is also summarized in Table 1. We also describe the NSGA-II implementation of this multi-objective problem and its components. We then describe the datasets and benchmark algorithms we used to evaluate the team formation problem. Finally, we explain the quantitative metrics to compare algorithms’ results.

Table 1. Notation.

Variable Definition
P Set of people
n Number of people
C Set of categorical attributes
U Set of numerical attributes
m Number of categorical attributes
l Number of numerical attributes
G Social network
p j jth person
d(pi,pj) Distance between persons i and j
T Set of teams
q Number of teams
k Number of members per team
P A set of team solutions (i.e., Population)
r Number of Team combinations (i.e., Population size)
F Pareto front

Definitions

Members, attributes, networks, and teams

We consider a set of participants P = {p1, p2, …, pn} with a set of categorical attributes C = {c1, c2, …, cm} and a set of numerical attributes U = {u1, u2, …, ul}. These individuals’ attributes have different scales and represent information about each person (e.g., age, gender, race, skill). Depending on the individuals’ information available, teams can have several attributes describing their qualities and composition. Each person has a value in each one of these attributes. We denote ci(pj) to obtain the value of the categorical attribute ci for the person j. Similarly, we use ui(pj) to obtain the value of the numerical attribute ui for the person j. Person j can be represented as a vector of these categorical and numerical attributes. Thus, we have the attributes of pj as (c1(pj), …, cm(pj), u1(pj), …, ul(pj)).

People are connected together in a social network modeled as an undirected and unweighted graph G. We define G = (P, E), where E represents the graph’s edges. Each node in G represents a person from P. We use person and node interchangeably throughout this paper. Two people are connected by an edge if they have collaborated in the past. In other words, if individuals i and j have worked together, then Gi,j = 1. Otherwise, Gi,j = 0.

Given this list of participants P connected in the network G, the goal is finding a set of teams T = {t1, t2, t3, …, tq}, where all members of P assemble q teams and belong to one team only. The optimization dual-problem can be formulated as minimizing the communication costs among team members and maximizing teams’ diversity levels. We now make these notions and describe each objective function.

Communication costs

Lappas et al. [57] focused on the importance of collaborations and familiarity between experts by considering the cost of their collaborations. According to this model, experts who collaborated together in the past are more likely to exchange information and ideas effectively than experts without prior collaborations. Based on experts’ prior collaborations, this model calculates the communication costs among team members to estimate their collaboration and familiarity levels. The goal of optimizing communication costs is to form teams with high familiarity levels. A literature review shows that communication costs are a highly used proxy for collaboration and familiarity among researchers [66].

In our setting, we use communication costs as a proxy for teams’ familiarity. Kargar and An [31] found the total sum of distances between team members to be a reasonable measure of communication costs, as it is more stable to changes in the network than other potential measures. Other alternatives for communication costs are the social network’s diameter (i.e., the largest shortest path between any two nodes in the network), and the minimum spanning tree (i.e., the minimum sum of the weights of a network’s edges) [57]. We also implemented this problem using these two definitions, and their results were similar to those obtained using the sum of distances. The results of the diameter’s implementation are available in S1 Fig and S1 Table in S1 File, and the results of the minimum spanning tree’s implementation are available in S2 Fig and S2 Table in S1 File.

We define the communication costs between two individuals pi and pj, denoted as d(pi, pj), as the shortest path length while traversing the edges of the graph G from one node to another. If pi and pj have collaborated in the past, they are at one-hop of distance. If pi and pj have not collaborated together but have a prior collaborator in common, they are separated by two-hops. Having common past collaborators within a team can promote familiarity based on “triadic closure” [67]. This mechanism posits that nodes are more likely to establish a new connection when they have a connection in common. Three-hops and 4-hops can follow the same principles based on “balance mechanisms” [67]. Individuals will tend to forge new connections with collaborators of their collaborators to seek consistency within their group. Therefore, using the total sum of distances in our objective function aims to search for teams that maximize the number of direct collaborations (i.e., one-hops), common connections (two-hops), and close connections (three-hops or higher). The lowest communication cost value is when all team members have collaborated with each other (i.e., they are directly connected), and the highest is when team members are not connected at all. In this implementation, if there is no path between pi and pj in G, we set the communication costs between them as the social network’s diameter.

We define the communication costs of a team t as the total sum of shortest path lengths between members, as it is more stable to changes in the network than other potential measures. We denote by Cc(t) the communication costs of team t, which has k members. Thus, we define the team t’s communication costs as:

Cct=i,jt,ijkd(pi,pj) (1)

The goal is to minimize the average sum of shortest path lengths across all assembled teams in the individuals’ network. Computing the sum of communication costs of a set of teams runs in O(n2) time.

Team diversity score

The second goal is to generate diverse teams with a broad array of backgrounds, traits, and skill repertoires. Diversity describes the distribution of differences among the members of a unit with respect to a common attribute [30]. Harrison and Klein [30] presented a framework suggesting that diversity is best conceptualized in three ways: separation, variety, and disparity. Separation refers to differences among team members in their lateral position on a continuum (e.g., value, attitude, belief). Variety refers to categorical differences among team members wherein the number of represented categories contribute to team diversity (e.g., gender, career, race). Finally, disparity represents differences in the concentration of valued assets or desirable resources (e.g., expertise, educational level, tenure). These metrics allow researchers to operationalize functional and demographic diversity in parallel and according to their theoretical conceptualizations [14].

In this implementation, we use variety metrics to assess teams’ diversity provided by C categorical variables, and disparity metrics to assess teams’ diversity provided by U numerical variables. To measure each team t’s variety metrics by its members’ categorical attributes C, we use the Blau index (Bt,ci) [30]. This index quantifies the probability that two team members randomly selected from the team would be in different categories. A low score means members fall into the same category, whereas a high score means members fall into different categories. We denote pcij is the proportion of members who fall into a particular category j in the categorical attribute ci. Given that the number of categories in ci is oci, where j=1,...,oci, the Blau Index’ formula for the team t is:

Bt,ci=1-jocipcij2 (2)

To measure each team t’s disparity metrics by its members’ numerical variables U, we use the coefficient of variation (CVt,ui) [30], which is defined as the ratio of the standard deviation to the mean of the attribute i, uiU. A low coefficient of variation score means that all team members t have similar levels of the attribute, whereas a high score means all team members t have different levels of the attribute. For a team t with members j = 1, 2, …, k, and with ui¯ as the team’s mean value of the attribute i, the formula is the following:

CVt,ui=jtk(ui(pj)-ui¯)2/k/ui¯ (3)

These two team diversity measures are useful because they do not change when the input data is scaled linearly, and they both tend to stay around the same values. Given that the team formation problem considers C categorical variables and U numerical variables, the diversity measures can be weighted to prioritize differences within a specific variable. The vector of weights W has |C| + |U| elements, where W=(wu1,...,wcm,wc1,wc2,...,wcl). Based on these measures, we aggregate diversity for different attributes into a single value. We define the team diversity score V of a team t as the weighted sum of the Blau indexes for all C categorical variables and the coefficient of variation for all U numerical variables. The formula is:

Vt=i=1mwui*CVt,ui+i=1lwci*Bt,ci (4)

Multi-objective problem statement

We formulate the problem as a multi-objective problem with the goal of finding a set of r team solutions P={T1,T2,T3,...,Tr}, where each T represents a potential solution with q teams. The decomposition of the assessment function into both objectives—minimizing communication costs and maximizing team diversity score—allows us to find multiple solutions that the single-objective approach cannot reach. As a result, we expect to find not a unique solution T but a set of solutions P for which there is not another feasible solution better in both objective functions. This set of solutions P is also known as a Pareto front, where (a) there exists no other set of solutions T′ with more diverse and connected teams, and (b) each solution Ti,iP is not superior to all the other solutions in P with respect to both diversity and communication costs objectives. Having this set of team solutions P makes it possible to assess each of them individually, so a team builder can choose the most appropriate teams possible to assemble for the given context and circumstances.

In sum, the team formation problem addressed in this paper is to find the Pareto front P of team solutions, where each solution T is composed by q teams (T = {t1, t2, t3, …, tq}). The dual objective is maximizing teams’ diversity based on the categorical attributes C and numerical attributes U and minimizing the communication cost based on G. We can model this problem as:

minCc(T)=t=1qCct;maxV(T)=t=1qVt;s.t.tPi=1q|ti|=ntT;t=1:q (5)

Since finding teams from a graph G while minimizing the sum of shortest path lengths and team allocation problems is proven to be an NP-hard problem [57, 68], this multi-objective problem is also an NP-hard problem.

NSGA-II implementation

Pareto fronts’ shapes provide helpful information about the degree of trade-off between different objectives and how much compromise is needed from some criteria to improve others. Determining the exact Pareto front for multi-objective combinatorial optimization problems is difficult since it is necessary to compute all the possible combinations to find the true Pareto front [63]. For this reason, the goal is to find an approximation of the true Pareto front using heuristic algorithms. A critical assumption for these algorithms is that the Pareto front is sufficiently populated. The quality of this approximation depends upon (1) the proximity of the points on the approximated front to the points on the true Pareto front; and (2) the diversity of the solutions on the approximated front, where more diversity is typically better. Although the true Pareto front is unknown, solutions that dominate others are close to the theoretical true Pareto front. Therefore, the diversity of the solutions will provide a more extensive range and granularity of the Pareto front.

Genetic algorithms (GA) are commonly used to find approximations of Pareto fronts [69]. By imitating evolution in nature, this method optimizes a population of initial solutions into better solutions through natural selection. Each solution is characterized as a chromosome (i.e., a vector of attributes), which can be mutated and altered in each iteration. The best solutions will endure after they mutate over time. Genetic algorithms are ideal for finding solutions for optimization problems in large and highly non-linear spaces [70].

The genetic algorithm starts from a population of randomly generated solutions evolving into new solutions through an iterative process. The population created in each iteration is also known as a generation. In each generation, the algorithm evaluates each population’s chromosome according to the objective function in the resolved optimization problem. The chromosomes with the highest scores are selected from the current generation and used to form a new generation. This process continues until a maximum number of iterations is achieved or by a threshold function defined for the solutions.

We implemented a genetic algorithm called Non-dominated Sorting Genetic Algorithm-II (NSGA-II) formulated by Deb et al. [71]. NSGA-II allows finding an approximation of the Pareto front, having different team solutions P that variate according to the communication costs and diversity score specified. The NSGA-II approach is based on sorting the populations into a hierarchy of sub-populations using Pareto dominance criteria. Then, chromosomes for the next iteration are selected according to the mentioned hierarchy. This elitist selection guarantees that potential good chromosomes are kept in the population, and the solution quality obtained does not decrease from one iteration to the next. The solutions are also ordered according to the similarity among their chromosomes, removing redundant ones to promote diversity in the Pareto front. As a result, NSGA-II can converge on a high-performing Pareto front after a few iterations. Previous work has shown that NSGA-II provides solutions with high levels of efficiency running in O(n2).

In this implementation, each population P contains r team solutions P={T1,T2,...,Tr}, and each chromosome represents a potential set of q teams Ti = {t1, t2, …, tq}. We use “chromosome” and “team solution” interchangeably throughout this paper. We characterize a chromosome as a vector of individuals partitioned into q parts to obtain the teams (Fig 2). As a result, each chromosome’s length equals the number of people n, representing q teams of size k (q*k = n). We adapted this algorithm to our specific diverse team formation problem, and we outline these steps in Algorithm 1.

Fig 2. Team chromosome representation.

Fig 2

Members only can be part of one team.

Algorithm 1: NSGA-II scheme pseudo-code

Input: Population size r, People P, Number of teams q, Number of Generations g

Output: Solutions P

P

for i ← 1 to r do

 Shuffle (P)

T ← Split(P, q)

 Add T to P

for i ← 1 to g do

 // Generate offspring

Children ← ∅

for j ← 1 to r do

  p1← RandomChoice (P)

  p2← RandomChoice (P)

  Add CrossoverAndMutation (p1, p2) to Children

 // Merge the parent population with the children population

P ← Merge (P, Children)

 // Evaluate communication cost and diversity score

 EvaluateCommunicationCost (P)

 EvaluateDiversity (P)

 // Perform non-dominated sort

F← FastNonDominatedSort (P)

 // Create a new population

P

 // Add fronts until we have the allowed population size

k ← 0

while Size (P) + Size (Fk) ≤ r do

  Add Fk to P

  kk + 1

 // Calculate the crowding distance of the last front

 CrowdingDistance (Fk)

 // Sort front’s solutions according to their crowding distance

 SortFronts (Fk)

 // Select final chromosomes and add them to the new population

 Δ ← r − Size (P)

FinalChromosomes ← SelectFinalChromosomes (Fk, Δ)

 Add FinalChromosomes to P

 // Update population for the next generation

PP

return P

Initialization

The algorithm starts by initializing a population of chromosomes P having teams assembled randomly. Its input parameters are the total number of chromosomes r to include in the population P, the list of people P, the number of teams q to form, and the number of iterations to perform g. Chromosomes are stored as two-dimensional arrays of shape (q, k), where q is the number of teams possible to assemble, and k is the number of members per team. Each chromosome is a potential solution to the diverse team formation problem, and the goal is to find a set of chromosomes with high levels of diversity and low communication costs. After the initial population is created, the algorithm creates the offspring and searches for the Pareto fronts iteratively until the maximum number of generations g is reached.

Crossover step

In each generation, the algorithm takes two random chromosomes (p1 and p2) from the existing population P and randomly selects q teams from this union. As a result, the algorithm will have a child chromosome with q teams. Since the child’s teams are randomly selected from two different chromosomes, individuals may get selected twice, coming from p1 and p2. The algorithm replaces repeated individuals with others who were not assigned to a team. It explores each member of the child chromosome and counts how many times an individual is part of a team. If an individual is counted more than once, this individual is randomly replaced by a missing member. At the end of this revision process, the algorithm will have the child chromosome with all the members of P assigned to one team. These random samplings provide sufficient mutation for the algorithm to introduce diversity into the population without adding another mutation step. We outline the proposed crossover method in Algorithm 2.

Union

After the crossover step, the algorithm combines the population P with its offspring, doubling the population’s size (i.e., 2r). The algorithm then calculates the diversity score V and communication costs CC of each chromosome of this union.

Fast non-dominated sort step

Next, the algorithm must select the best r chromosomes from this union of size 2r. To find this set, the algorithm performs a non-dominated sorting among all the existing chromosomes from P. The goal is to identify solutions that perform better than others and classify them according to their performance in different Pareto fronts F. The algorithm first checks the dominance relationships among all the chromosomes. Given two chromosomes, T and T′, T dominates T′ if and only if Cc(T)≤Cc(T′) and V(T)≥V(T′) with at least one strict inequality. In other words, T is at least as good as T′ for all objectives and strictly better for at least one. This dominance relation is denoted as TT′. If one of the objectives of T is not better than T′, and it cannot be improved in value without degrading some of the other objective values, then T is non-dominated by T′. One example of a non-dominated solution is T having higher diversity scores but higher communication costs than T′. In that non-dominance case, either T and T′ are feasible solutions for the next generation.

Once the algorithm maps all the chromosomes’ dominance relationships, it creates a first Pareto front of solutions consisting of all the non-dominated solutions (F1). This set is also denominated as the Pareto optimal. Then, the algorithm creates a second front of Pareto optimal solutions (F2) that were disregarded in the first front, and so on. As a result, the algorithm sorts the population’s chromosomes into a hierarchy of sub-populations. The sort keeps finding successive Pareto fronts until all chromosomes are assigned to a Pareto front.

New population

The algorithm then selects the best r chromosomes for the next generation. At a given time, there are 2r chromosomes sorted in the hierarchical Pareto fronts F. The algorithm creates the new population P adding the chromosomes stored in the Pareto fronts. If the total size of the first Pareto front is smaller than r, then the algorithm adds all the chromosomes of this front to P. Then, the algorithm adds the remaining solutions for the new population from the subsequent non-dominated fronts. The algorithm continues this procedure until it can not add more fronts to P.

Crowding distance

The algorithm must add chromosomes to the new population until there are exactly r chromosomes. If the last selected non-dominated Pareto front Fk has more chromosomes than the allowed to add to P, the algorithm must choose a smaller set from Fk to complete the r chromosomes. Let δ=r-Size(P), the number of missing chromosomes to complete r. The algorithm identifies the best δ chromosomes from this last front Fk by calculating the crowding distance among the chromosomes. This metric determines how similar the chromosomes are in terms of performance in the multi-objective problem. After calculating this distance, the algorithm ranks the chromosomes according to their distances and eliminates chromosomes that perform similarly to other chromosomes. This procedure keeps a broader front of solutions and removes redundant chromosomes. Then, the δ best chromosomes from Fk are added to P. As a result, P counts with the r best chromosomes and becomes the parent of the next generation, starting a new iteration.

Output

After the optimization runs through the previously specified number of generations g, the algorithm returns an approximation of the Pareto front P having r team solutions.

Algorithm 2: Crossover Function

Input: Parent p1, Parent p2, People P, Number of Teams q

Output: Children

p ← Concatenate (p1, p2)

Children ← AssembleRandomTeams(p, q)

MissingMembers ← Set (P) - Set (Children)

Counted ← ∅

for Child in Children do

if Child in Counted then

  NewMember ← SelectRandomMember (MissingMembers)

  Replace ChildNewMember

  Remove NewMember from MissingMembers

else

  Add Child to Counted

end

end

return Children

Data

In this section, we evaluate the proposed algorithm for our team formation problem using three real-world datasets. The data sources are MyDreamTeam (a team formation platform), Bibsonomy (a social bookmarking site), and GHTorrent (a GitHub repository database). Using these datasets to simulate teams for this team formation problem illustrates the effectiveness of our framework in real scenarios. We show summary statistics from these datasets in Table 2. The resulting data and the scripts to pre-process the raw data are available at http://nusoniclab.github.io/.

Table 2. Description of the MyDreamTeam, Bibsonomy, and GHTorrent datasets.

Case Description TS N E D MSD De. Cent.
MDT Case A An undergraduate course at an university in the US 5 55 130 7 3.03 0.06 0.14
MDT Case B A graduate course at an university in the US 6 61 518 4 2.23 0.14 0.42
MDT Case C An MBA course at an university in the US 3 65 87 8 3.51 0.03 0.06
Bibsonomy Case A Papers published at Nature 5 40 398 3 1.39 0.26 0.22
Bibsonomy Case B Papers published at Science 5 100 2,238 4 1.14 0.23 0.24
Bibsonomy Case C Papers published at Physica A. 5 86 160 4 1.58 0.02 0.07
GHTorrent Case A Repositories with Python as the main language 5 57 318 4 1.64 0.10 0.12
GHTorrent Case B Repositories with Java as the main language 5 100 4,683 2 1.05 0.47 0.03
GHTorrent Case C Repositories with Ruby as the main language 5 55 642 3 1.17 0.43 0.22

We include TS: team size, N: the number of individuals, E: the number of relationships among individuals, D: the collaboration network’s diameter, MSD: individuals’ mean short distance, De.: the collaboration network’s density, and Cent.: centralization score using degree centrality.

MyDreamTeam dataset

We evaluate our proposed algorithm using data from real team formation cases. We extracted this dataset from the My Dream Team Builder [33], a recommender system to help individuals self-assemble teams. This dataset contains cases of participants self-assembling their teams. Cases date from 2014 to 2020. On this recommender system, participants create profiles, search for teammates, and send invitations to form teams. The cases consist of classes from universities in the United States. The dataset includes participants’ traits, demographics, and social networks, which they reported in an initial survey. We selected three cases to test our algorithm: an undergraduate course, a graduate course, and an MBA course. Participants used the system to assemble teams for small group discussions.

Permission to collect data from participants was approved by Northwestern University Institutional Review Board (#STU00078513). All applicable institutional and governmental regulations concerning the ethical use of human subjects were followed during this research. Electronic consent was obtained from study participants via an online survey instrument. Participants were asked for their consent to use data collected through My Dream Team Builder for research purposes. We hashed users’ identifiers to create a de-identified dataset.

bibsonomy

The second dataset is extracted from bibsonomy [34], a social-bookmarking and publication-sharing system. We chose bibsonomy since prior team formation papers tested their algorithms using this database [58]. This dataset is administered by the Knowledge and Data Engineering Group, University of Kassel. The bibsonomy dataset is available under a license agreement, and it can be requested at https://www.kde.cs.uni-kassel.de/wp-content/uploads/bibsonomy/. This dataset contains a large number of computer science related publications. Each publication is written by a group of authors. The bibsonomy website is visited by many users who use tags to annotate the publications. Following the procedure described by Anagnostopoulos et al. [58], we used the tags associated with each author’s papers to represent their skills. Each author’s skill represents the number of papers published with their respective tag. We selected three journals related to social network analysis to test our algorithm: “Nature”, “Science”, and “Physica A: Statistical Mechanics and its Applications.” We counted the frequency of the tags in each of these journals and selected some popular tags related to our study. For the first two journals, we selected papers that included the tags ‘network’, ‘social network’, and ‘small world.’ Then, we identified the authors of these articles, created the co-authorship network, and selected authors from the largest component. Similarly, we did this procedure for the third journal using the tags ‘network’, ‘graph’, ‘model’, and ‘system.’ We hashed the authors’ names to create a de-identified dataset.

GHTorrent

We used GitHub data provided by the GHTorrent project [35], an offline mirror of the data offered through the GitHub API. This dataset can be downloaded at https://ghtorrent.org/downloads.html. The GHTorrent dataset covers a broad range of development activities on Github, including repositories, pull requests, and users. We downloaded the dataset dump “06/01/2019” to build our testing dataset. We filtered users who contributed between 40 and 80 projects to keep median users in our analysis. Following an approach similar to the bibsonomy dataset, we used programming languages associated with each user’s contributed repositories to represent users’ skills. Each user’s skill represents the number of contributed projects written in a specific language. Since repositories can have files in multiple languages, we selected repositories’ most used language as the repository’s language. We selected three of the most popular languages in this dataset: Java, Python, and Ruby. Then, we identified the users of these repositories and created the collaboration network. In this example, users have a tie if they contributed to the same repository at least two times. Finally, we selected users from the largest component. We hashed the authors’ names to create a de-identified dataset.

Evaluation

We compare the proposed algorithm for the team formation problem (denoted as NSGA-II) against three well-known multi-objective optimization methods used for benchmark purposes [62, 72]:

Pareto Local Search (PLS) method

This iterative algorithm starts with a set of random solutions as the initial population and explores each solution’s neighbors [73, 74]. The algorithm updates the population based on Pareto dominance: it will add non-dominated neighbors to the population and remove existing solutions that are dominated by the newly added solutions. Once the neighborhood of a solution has been fully explored, the solution is marked as explored. The algorithm iteratively explores new solutions as they are added to the population until no better solutions are found. After all the solutions are explored, and no more non-dominated solutions can be discovered, the algorithm stops. We implemented the version proposed by Zihayat et al. [72] for combinational problems. In this implementation, a solution’s neighbors are all the possible team combinations from the solution with two members swapping teams. Since PLS does not depend on a fixed number of generations, we only run one iteration of this algorithm to compare its results with the other methods. Given n individuals, and that the algorithm will explore (n2) neighbors of each solution, the computational complexity of this implementation is O(n3) in the best-case scenario.

Strength Pareto Evolutionary Algorithm 2 (SPEA-2)

Like NSGA-II, this algorithm is based on elitist selection and dominance criteria [75]. Instead of creating different Pareto fronts, SPEA-2 keeps the set with the best solutions found in each iteration called “archive,” which is separated from the population. The algorithm starts with random population solutions and an empty archive. Then, it calculates a fitness value for each solution based on (a) the number of solutions it dominates (i.e., strength), (b) the number of solutions by which it is dominated by the current population (i.e., raw fitness), and (c) its distance with other solutions (i.e., density value). The best solutions will be copied to the archive. After initiating the first population, the goal is to identify non-dominated solutions for the next generation. Based on the fitness values, the algorithm performs binary tournament, crossover, and mutation steps with the solutions from the current population and archive. These new solutions will constitute the next population. After these processes, the algorithm checks how many non-dominated solutions result from the union of the current population and archive. If the number of non-dominated solutions is less than the archive’s size, the archive will include some dominated solutions from the union. The algorithm selects dominated solutions based on their fitness values. If the number of non-dominated solutions is higher than the archive’s size, the algorithm removes redundant solutions based on their nearest neighbor Euclidean distance. The next iteration will create a new generation based on this updated archive. We implemented the version proposed by Zitzler et al. [75]. We used the same number of generations from the NSGA-II testing and set the archive’s size to equal the population’s size. In the best-case scenario, the computational complexity of this algorithm is O(M2logM) where M is the sum of the population size (n) and archive size (n′).

Hybrid Particle Swarm Optimization (HPSO) method

This algorithm combines the steps of particle swarm optimization algorithms (PSO) and genetic algorithms (GA) [76]. In its original version, PSO starts with a population of candidate solutions (called particles) and moves them around in the search space over the particle’s position and velocity. Each particle’s movement is influenced by its local best-known position, but is also guided toward the global best-known positions in the search space. In each iteration, the algorithm updates particles’ positions based on their velocity. After a few iterations, the algorithm provides solutions that are approximations of local optima and global optima. Since the PSO’s original formulation only operates in continuous optimization problems, we require a version that can handle combinational optimization problems. Moreover, PSO operates with a global optimum that does not exist in Pareto front problems. Zhang et al. [76] proposed a hybrid version that replaces the PSO’s particle position and velocity update formulas with the genetic algorithm’s crossover and mutation operations. In a nutshell, the HPSO algorithm iteratively examines each particle and (a) applies the crossover step with a random non-dominated solution found by the particle, (b) applies the crossover step with a random non-dominated solution known from all the population, (c) and performs the mutation step. If a resultant solution is better than the original, then the solution is updated. If a particle knows two or more non-dominated solutions, it will choose a random non-dominated solution as the best local particle. Similarly, if the population knows more than one non-dominated solution, it will select a random non-dominated solution as the best global particle. The running time of this algorithm is expected to be polynomial since it will check the n solutions and run the crossover operation two times and the mutation operation once. As a result, the computational complexity is O(n2) in the best-case scenario.

We also compared the teams assembled by these four multi-objective algorithms with randomly assigned teams. Since the MyDreamTeam dataset already included fixed-size teams, we also computed the real teams’ diversity scores and communication costs.

Metrics

We computed the following quantitative metrics to evaluate the quality, quantity, and running time of the algorithms’ solutions. These indicators map the final solutions to a number that indicates one or several aspects of the solution. We chose these metrics based on the literature review by Li et al. [77].

Hypervolume (HV)

This metric evaluates the total size of the objective space dominated by the algorithm’s solutions with respect to a reference point. It can measure how close solutions are to the true Pareto front and how evenly spread the solutions are in the objective space. Algorithm A will have higher hypervolume scores than algorithm B if algorithm A’s solutions dominate algorithm B’s solutions. In this context, higher hypervolume scores show that team combinations with higher levels of diversity and familiarity can be found. If the algorithm A finds team combinations with higher diversity scores and/or lower communication costs than algorithm B, the algorithm A’s hypervolume will be higher than the algorithm B’s hypervolume. The larger the HV value, the better the diversity and distribution of the team combinations. The HV of an algorithm A can be formulated as:

HV(A)=λ(aAx|axr) (6)

where r denotes the reference point, and λ indicates a measure to subsets of n-dimensional Euclidean space (i.e., Lebesgue measure). In our case, the hypervolume is the area of the rectangles formed by the solutions and a two-dimensions reference point.

Unique Non-dominated Front Ratio (UNFR)

This metric quantifies the contribution of each algorithm to the combined non-dominated front of all the algorithms. In this context, if algorithm A has a higher UNFR value than algorithm B, the former found team combinations with higher diversity and/or lower diversity scores than the latter. Let Aunf be the unique non-dominated front of a given algorithm A, then this metric is defined as:

UNFR(A)=|aAunf,rRunf:ra||Runf| (7)

where Runf is the set of unique non-dominated solutions of the collections of all solutions produced by the algorithms. The UNFR value ranges from 0 to 1. An algorithm with a high UNFR value means that it contributed to many unique non-dominated solutions from all the non-dominated solutions found. In contrast, a value close to zero means that the algorithm provided a few unique non-dominated solutions to the final set.

Computational complexity

Lastly, we evaluated these algorithms’ computational complexity as a function of the input size. In this context, if algorithm A has a lower running time than algorithm B, the former can find team combinations from a pool of participants faster than the latter. Since some algorithms’ running time can increase exponentially, this metric is relevant to measure how scalable and efficient the algorithm is when forming teams with large participant pools. We compared the algorithms’ running times using different numbers of users from the GHTorrent “Java” and Bibsonomy “Science” datasets.

Results

We ran the algorithms’ evaluations for 50 generations with a population size of 50 chromosomes. We implemented these algorithms in Python 3.6.2. and performed the experiments on a server with a 2.60 GHz Intel(R) Xeon(R) CPU and 16GB of RAM. The algorithms’ implementations and detailed results are available at http://nusoniclab.github.io/ for consultation. Table 2 shows the statistical data of the datasets, including the team size, the number of available individuals, the number of relationships, the diameter of the network, individuals’ mean short distance, and networks’ centralization.

Fig 3 shows the approximation of the Pareto front found by each algorithm in each dataset. The x-axis represents teams’ total communication costs. Lower scores on this axis represent solutions with lower communication costs (i.e., teams internally more connected). The y-axis represents the total teams’ diversity score of the solutions. Higher scores in that axis represent solutions with more diverse teams. As the results show, the NSGA-II implementation outperforms the benchmark algorithms in most of the tested datasets. NSGA-II found non-dominated solutions with high diversity values and low communication costs across all these databases. HPSO also contributed with non-dominated solutions to the final set of solutions. In particular, the plots show that HPSO was better at finding non-dominated solutions when setting a balanced trade-off between communication costs and diversity. Following NSGA-II and HPSO, PLS solutions were close and concentrated in certain regions of the team formation space. This concentration indicates that PLS tended to converge on certain non-dominated solutions, dismissing other potential team combinations that may not have been non-dominated in the first iterations. SPEA-2 results were worse than the other algorithms despite employing the same representation and operations. Overall, NSGA-II was better at finding solutions in the extremes of the approximate Pareto front, offering more variety of non-dominated solutions. It provided more alternatives compared to PLS, HPSO, and SPEA-2. Therefore, the NSGA-II implementation provides a spectrum of team solutions that team builders can explore and choose.

Fig 3. Results of NSGA-II, SPEA-2, PLS, HPSO, and random assignment.

Fig 3

The NSGA-II implementation provided more diverse team solutions with high diversity scores and low communication costs than the solutions provided by PLS, HPSO, and SPEA-2. (a) MyDreamTeam: Experiment A. (b) MyDreamTeam: Experiment B. (c) MyDreamTeam: Experiment C. (d) Bibsonomy: Nature papers. (e) Bibsonomy: Science papers. (f) Bibsonomy: Physica A. papers. (g) GHTorrent: Python repositories. (h) GHTorrent: Java repositories. (i) GHTorrent: Ruby repositories.

The results show that the NSGA-II algorithm achieved the largest hypervolume values on 6 of 9 datasets and the second-highest value for the other three datasets (Table 3). In other words, NSGA-II frequently found more team combinations with higher diversity levels and lower communication costs than the other algorithms. NSGA-II’s high hypervolume values can be explained by its crowding distance step, which helped the algorithm find non-dominated solutions located on the extremes of the Pareto front. Since PLS and HPSO did not establish any criteria to avoid redundant solutions, their solutions resulted in a set of non-dominated solutions concentrated in certain areas. Therefore, the set of team combinations provided by NSGA-II frequently dominated the ones provided by the other algorithms.

Table 3. Hypervolume and unique non-dominated front ratio values for the five methods across the datasets.

Hypervolume Unique Non-dominated Front Ratio
HPSO NSGA-II PLS Random SPEA-2 HPSO NSGA-II PLS Random SPEA-2
MDT Case A 29,755 29,696 28,536 26,304 27,292 0.67 0.33 0 0 0
MDT Case B 30,298 29,972 29,151 28,506 28,593 0.86 0.14 0 0 0
MDT Case C 92,643 95,205 89,531 84,546 83,331 0.09 0.91 0 0 0
Bibsonomy Nature 368,697 364,369 339,493 345,935 322,730 0.29 0.71 0 0 0
Bibsonomy Science 684,658 721,970 674,118 631,915 529,396 0.00 1.00 0 0 0
Bibsonomy Physics A 386,482 394,852 356,903 321,826 294,007 0.44 0.33 0.22 0 0
GHTorrent Case Python 138,632 139,959 121,850 119,406 109,592 0.30 0.70 0 0 0
GHTorrent Case Java 15,572 15,614 14,967 14,923 14,886 0.33 0.67 0 0 0
GHTorrent Case Ruby 99,556 103,913 95,745 92,445 91,108 0.21 0.79 0 0 0

The best results are marked in bold.

The NSGA-II implementation also scored the highest unique non-dominated front ratio (UNFR) values on 6 of 9 datasets. In other words, NSGA-II frequently provided more non-dominated team combinations that the other algorithms could not find. A possible explanation for the NSGA-II second-place in the other cases is the low density in the collaboration network. Pools with few prior connections among individuals will reduce the number of possible highly-connected team combinations, making the heuristic search ineffective. In contrast, HPSO and PLS performed more crossover and mutation operations than NSGA-II. Executing these operations multiple times allowed HPSO and PLS to examine more team combinations and increase their likelihood of finding specific team combinations with low communication scores.

HPSO achieved the second-highest hypervolume and UNFR values. It benefited from the non-dominated solutions in the middle of the approximate Pareto front, which scored high diversity levels. These non-dominant solutions outperformed other algorithms and increased the area created by its approximate Pareto Front. SPEA-2 and PLS converged to a few solutions, covering a smaller area than the NSGA-II and HPSO solutions. Overall, NSGA-II found more non-dominated solutions across these two objectives and provided solutions with higher variance in communication cost values.

The large variance in both diversity and familiarity shows that the NSGA-II algorithm found more non-dominated solutions than the other algorithms, which is desirable to find the true Pareto front. The crowding distance step of NSGA-II allowed the algorithm to keep a broader range of non-dominated solutions. Plus, the algorithm kept secondary solutions in different layers that could have originated non-dominated solutions in later iterations. As the algorithm continues creating new generations, dominated solutions can still be considered to find other potential solutions. Moreover, NSGA-II could still identify non-dominated solutions in the middle of the trade-off. In contrast, the low variance of the other algorithms shows that they were likely to converge on a specific set of non-dominated solutions and to a specific trade-off. These algorithms did not include operations to diversify their current populations or remove redundant non-dominated solutions. Therefore, these algorithms could lack diverse solutions that reside in the extremes of the Pareto front.

Fig 4 presents the run time of all algorithms implemented. Fig 5 shows how algorithms’ running time as a function of the number of available individuals. The results show that the NSGA-II implementation performs better than PLS and HPSO as the participant pool increases. PLS required more time to explore solutions’ neighborhoods until all possible combinations were exhausted. In the case of HPSO, the two crossovers and one mutation step performed for each solution made the algorithm’s operation three times longer than NSGA-II since the latter only performs one crossover step. Although HPSO took longer than NSGA-II, both worked in polynomial time (O(n2)). Our results suggest NSGA-II required less than one-third of the time that PLS and HPSO took to provide similar results. Therefore, using NSGA-II is highly encouraged to find solutions efficiently as the input size increases. SPEA-2 did not find better solutions than PLS or NSGA-II, but its results converged faster than the NSGA-II and PLS methods.

Fig 4. Running time results for the NSGA-II, SPEA-2, PLS, HPSO, and random assignment.

Fig 4

All metrics are shown on a logarithmic scale.

Fig 5. Comparison of algorithms’ time complexity by varying the number of users.

Fig 5

(a) Bibsonomy Science dataset. (b) GHTorrent Java.

Lastly, we calculated the frequency of direct contacts (1-hop), shared contacts (2-hops), 3-hops, and more in the assembled teams to understand the distances among team members (See S3 Table in S1 File). The results suggest that the vast majority of the members were connected with others through one intermediate (∼31%), followed by members who were directly connected (∼30%). These numbers show that the resultant teams were highly connected in general, and members connected through many hops were not representative.

Discussion

Forming teams is a challenging task, especially when the goal is to bring a balance between diversity and members’ teammate preferences. While prior work has focused on looking for the best team possible [37], the contribution of this paper is searching for balanced team combinations that distribute skills and connections equitably. Moreover, assembling teams that include both diversity in their members’ contributions and prior relationships between members becomes an ultimate challenge to guarantee their success [78, 79]. In this work, we consider the problem of creating teams that are both diverse and highly connected teams using a computational approach. We implemented this team formation problem using a genetic algorithm that provides different team combinations according to teams’ diversity and communication costs. The results indicate that diverse and highly connected teams can be assembled efficiently and quickly through this genetic approach. In the following subsections, we elaborate on the implications of this work and the potential applications.

This work shows the benefits of using computational approaches to assemble multiple teams that distribute members’ skills among different groups and consider their prior relationships. Prior work mainly emphasizes finding the best team possible from a social pool (e.g., expert-team, all-star team) [36, 57, 58]. However, searching for multiple teams is also relevant in learning and organizational situations, such as forming student teams, forming operational teams within business units, or conducting scientific research. Because of the strong emphasis on assembling high-performing teams, algorithms that bring the best members together can create greater segregation within these social groups at the expense of others who are less skilled [80]. The concentration of skills and expertise in a small number of teams hinders the emergence of other teams having similar opportunities and social connections. As this paper demonstrates, the advanced computational infrastructure and the use of big data provide new opportunities to re-imagine several team combinations that individuals could not systematically and easily explore [37, 44]. Rather than creating teams using manual strategies or intuition, team builders can use algorithms to tailor different team combinations that optimize diversity and familiarity simultaneously. Thus, implementations like the one presented in this paper allow team builders to create heterogeneous and diverse teams without sacrificing familiarity among team members, which is essential for successful collaborations [81].

Another contribution of this work is finding similar team combinations that individuals would assemble but with enhanced diversity levels. As prior studies have found, people tend to form teams with competent individuals and those who are familiar with them, enhancing the likelihood of satisfaction and commitment with the team [28, 82]. This fact is demonstrated in the MyDreamTeam dataset by comparing the lower communication costs of self-assembled teams and the higher communication costs of randomly generated teams. The proposed algorithm found team combinations with lower communication costs than self-assembled teams, suggesting that people have some intuition in forming well-connected teams. However, they lack reliable knowledge of higher-order connections among themselves. A possible explanation of this difference found by the algorithm is the tremendous challenge for individuals to discover and take advantage of indirect connections, such as shared contacts or shared past collaborators. Whether individuals assemble their teams or team builders design them, considering team members’ indirect connections is not an easy task since indirect connections are not highly visible. In contrast, our algorithm excels in considering the broader social network structure given the global view of relationships between members. By using this algorithmic approach, individuals and managers can be more conscious of potential diverse teammates through their current relationships. Even if two team members do not know each other directly, teaming up with a shared “friend-of-a-friend” or indirect connections can potentially promote familiarity and psychological safety in teams [8385].

Furthermore, we found that MyDreamTeam self-assembled teams were less diverse than the teams randomly generated by the algorithms. This tendency driven by homophily is consistent with prior literature, indicating that people prefer to team up with others who share similar characteristics [65]. Formulating this team formation problem provides new opportunities to boost team diversity over self-assembled teams while still considering high familiarity among team members. One main advantage of forming teams in this fashion is reducing individuals’ biases. Since people naturally draw toward forming teams with similar individuals, an algorithm like the proposed one can augment people’s decision-making process. Instead of connections driven by individuals’ preferences, the algorithm can enact collective coordination by curating better team combinations that could satisfy individuals’ expectations. This multi-objective approach can allow people to find feasible solutions that increase diversity without compromising familiarity in the team.

Implications

This work provides theoretical implications for team research. In particular, the use of computational mechanisms to support the team formation processes. Literature has characterized team formation centered on behavioral mechanisms, where teams can be assembled by internal or external forces and based on similarity, familiarity, and competence [28, 86]. By formulating and implementing this multi-objective optimization problem, we found diverse and connected team combinations that individuals could not have foreseen. This work allows team scholars to reflect on the role of technologies in enabling new organizational structures among individuals and organizations, which could lead to new theories of team formation and the introduction of technologies [3840].

Practical implications of this study contribute to several communities invested in increasing team diversity: managers assembling effective and diverse teams, instructors composing more diverse student teams, companies forming heterogeneous groups from different business units, space agencies such as NASA forming composing space crews for long duration space exploration to Mars, and researchers investigating the use of algorithms for organizing scientific teams. Expanding the use of this algorithm to broader audiences can provide new benefits for groups that seek to embrace diversity and keep high familiarity levels. Furthermore, software developers and designers can use this study’s implications for new procedures and guidelines for artificial intelligence in organizing workers. Finally, this work provides more computational approaches to enrich team formation processes [45, 87]. Since team builders cannot solve this problem quickly by manually checking each team combination, algorithms can automatize this task by bringing together members who possess existing social connections and, at the same time, have different backgrounds, characteristics, and expertise levels [41, 42]. We expect this work will assist in forming heterogeneous teams by considering diversity and social networks.

Another quality of this approach is adding more objectives to the team formation problem. For example, team builders could minimize other objective functions such as geographical distance among participants, personnel costs, or availability constraints. Likewise, this multi-objective problem can accommodate members’ traits when diversifying is not desirable. As some prior meta-reviews indicate [14, 88], having a team with similar individuals may be desirable for low-difficulty tasks or when efficiency (rather than creativity) is the goal. Furthermore, it may be desirable for some traits such as personality or expertise to be similar rather than diverse [89]. This team formation problem can add another objective function that minimizes teams’ diversity in some traits using the metrics defined by Harrison and Klein [30]. Therefore, one potential use of this algorithm is to maximize diversity in some members’ attributes while minimizing diversity in other attributes.

Given this multi-objective approach’s flexible trade-off, which solution should team builders consider from the Pareto front? Incorporating other metrics (e.g., individual performance, team cohesion, members’ location) could help team builders select one specific team combination.

Limitations and future work

It is important to acknowledge the limitations of this paper. First, the measures for diversity and communication costs were scaled specifically to each unique network and cannot be compared across different sets of participants. Second, the diversity measure is an aggregate of multiple diversity metrics for each attribute sampled; thus, it is difficult to assign any real meaning to the diversity metric apart from relative differences with the same network. Future implementations should consider how different diversity measures can be analyzed separately and according to the specific pool of participants. These might also weigh diversity on various dimensions or operationalize diversity metrics as different objective functions in the optimization problem. Third, forming scientific teams and software teams is more complex in reality: new members can be added over time, some specialization is required, not all of these teams share the same objectives, sizes, or restrictions, and diversity may be beneficial for only goals. We believe using the last two datasets should not be a concern because we use them only to test the algorithms’ efficiency and results. This team formation algorithm can guide the formation of real scientific and software teams by finding more diverse and connected combinations. Fourth, we do not provide specific recommendations for demographic or functional diversity attributes. Prior studies have shown how the effects of diversity on team performance are mediated by contextual factors and team processes [14]. Team builders who want to administer this algorithm should reflect and decide on adding demographic and cognitive variables according to their organizational goals and particular context. Fifth, collecting social network data could be a big task for team builders. Assessing people’s relationships can be done by conducting surveys, examining communication networks, or tracing digital data [90]. Another potential strategy to build individuals’ social networks is asking about their teammate preferences. The algorithm could find possible diverse team combinations based on individuals’ responses [91]. Lastly, it is not possible to guarantee that the performance of the teams assembled by this algorithm will be better than other team formation strategies. Prior studies have shown mixed results for the direct effect of diversity on team performance in all contexts [14], as well as the advantage of using algorithmic approaches for team formation [92]. Other research has also shown that when individuals lack agency to self-assemble teams, they are less committed to their group [93, 94]. Future work should consider using this algorithm to assemble real groups and evaluate how well they perform compared to teams assigned randomly or by a manager.

Future work should add new restrictions to the multi-objective function, such as considering specific task roles in the teams, adding leaders to each team, or excluding certain team combinations in which some individuals do not want to work together. Using weighted networks could also provide more nuanced information about the strength of people’s social relationships. One potential application is distinguishing individuals who have frequent interactions from those who barely speak to each other [95]. One example of potential areas of improvement is developing an automatic tuning for the weights assigned for each diversity attribute given a specific population. If the algorithm explores people’s categorical and numerical attributes before conducting the team formation process, it could identify the attributes that have more variation and those that are scarce among individuals. Then, the algorithm could define the importance of each diversity attribute in the objective function. Lastly, the algorithm could be implemented as a web platform to provide more opportunities for managers, instructors, or researchers to assemble teams.

Conclusion

This work addresses the problem of assembling teams from a social network that maximizes both diversity and familiarity. We formulated a multi-objective function for this problem and implemented a genetic algorithm to find well-connected diverse teams. In a thorough experimental evaluation, we evaluated the performance of our proposed algorithm and compared it against baseline approaches. We discussed the potential role of algorithms in augmenting team composition and helping team builders. In particular, computational approaches can be used to form teams that consider indirect connections and recommend combinations with higher diversity scores. As algorithms can discover more feasible team combinations than humans, team builders’ decisions can become more structured, systematic, and comprehensive.

Supporting information

S1 File. Supporting figures and tables.

S1 Fig: Simulations using the Diameter metric. S2 Fig: Simulations using the Minimum Spanning Tree (MST) metric. S1 Table: Diameter Case. S2 Table: Minimum Spanning Tree Case. S3 Table: Team combinations’ average proportion of hops.

(PDF)

Data Availability

We have deposited a minimal data set in the GitHub repository https://nusoniclab.github.io/. This repository contains the following files: (1) the pre-processed and de-identified data used in this study, (2) the Python scripts to pre-process the original datasets, (3) the Python scripts that run the proposed algorithm and benchmark algorithms, including their plots and quantitative metrics. The pre-processed data in this repository constitutes the minimal dataset required to replicate all study findings reported in this article. The first dataset (MyDreamTeam) is administered by the SONIC Research Group, Northwestern University. We deposited a de-identified and minimal dataset from MyDreamTeam in the Github Repository to replicate the findings reported in this article. Because of the sensitive nature of some of the variables collected, Northwestern University Institutional Review Board (IRB)-approved protocol does not permit individual-level data to be made unrestricted and publicly available. Researchers interested in obtaining a larger sample of this restricted and de-identified individual-level data should contact the authors to inquire about receiving an IRB-approved institutional data sharing agreement. The second dataset used (bibsonomy) is administered by the Knowledge and Data Engineering Group, University of Kassel. This dataset is available under a license agreement, and it can be requested at https://www.kde.cs.uni-kassel.de/wp-content/uploads/bibsonomy/. While we do not maintain this dataset, we provide the scripts to generate the pre-processed datasets used in this study. The third dataset (GHTorrent) is administered by Georgios Gousios. The dataset is freely and publicly available at https://ghtorrent.org/. While we do not maintain these datasets, we provide the scripts to generate the pre-processed datasets used in this study.

Funding Statement

This study was supported by the National Institute of Health (1R01GM112938-01, 1R01GM137410-01), the National Aeronautics and Space Administration (80NSSC21K0925), and the National Science Foundation (SMA-1856090) through grants awarded to NC. This study was also supported by the Directorate for Social, Behavioral and Economic Sciences (SES-2021117) and Microsoft Research (2020 Microsoft Research Dissertation Grant) through grants awarded to DG. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Ng ESW, Burke RJ. Person–organization fit and the war for talent: does diversity management make a difference? The International Journal of Human Resource Management. 2005;16(7):1195–1210. doi: 10.1080/09585190500144038 [DOI] [Google Scholar]
  • 2. Hunt V, Layton D, Prince S. Diversity matters. McKinsey & Company. 2015;1(1):15–29. [Google Scholar]
  • 3. Jackson SE, Joshi A. Work team diversity. In: APA handbook of industrial and organizational psychology, Vol 1: Building and developing the organization. American Psychological Association; 2011. p. 651–686. [Google Scholar]
  • 4. Barak MEM, Travis DJ. Socioeconomic trends: Broadening the diversity ecosystem. In: The Oxford handbook of diversity and work. Oxford University Press; 2012. p. 393. [Google Scholar]
  • 5. Mathieu JE, Hollenbeck JR, van Knippenberg D, Ilgen DR. A century of work teams in the Journal of Applied Psychology. Journal of applied psychology. 2017;102(3):452. doi: 10.1037/apl0000128 [DOI] [PubMed] [Google Scholar]
  • 6. Williams KY, O’Reilly CA III. Demography and Diversity in Organizations: A Review of 40 Years of Research. Research in organizational behavior. 1998;20:77–140. [Google Scholar]
  • 7. Van Knippenberg D, De Dreu CK, Homan AC. Work group diversity and group performance: an integrative model and research agenda. Journal of applied psychology. 2004;89(6):1008. doi: 10.1037/0021-9010.89.6.1008 [DOI] [PubMed] [Google Scholar]
  • 8. Page SE. The diversity bonus: How great teams pay off in the knowledge economy. Princeton University Press; 2019. [Google Scholar]
  • 9.Vasilescu B, Posnett D, Ray B, van den Brand MGJ, Serebrenik A, Devanbu P, et al. Gender and Tenure Diversity in GitHub Teams. In: Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. CHI’15. New York, NY, USA: Association for Computing Machinery; 2015. p. 3789–3798. Available from: 10.1145/2702123.2702549. [DOI]
  • 10. Woolley AW, Aggarwal I, Malone TW. Collective Intelligence and Group Performance. Current Directions in Psychological Science. 2015;24(6):420–424. doi: 10.1177/0963721415599543 [DOI] [Google Scholar]
  • 11. Díaz-García C, González-Moreno A, Sáez-Martínez FJ. Gender diversity within R&D teams: Its impact on radicalness of innovation. Innovation. 2013;15(2):149–160. doi: 10.5172/impp.2013.15.2.149 [DOI] [Google Scholar]
  • 12. Andrevski G, Richard OC, Shaw JD, Ferrier WJ. Racial Diversity and Firm Performance: The Mediating Role of Competitive Intensity. Journal of Management. 2014;40(3):820–844. doi: 10.1177/0149206311424318 [DOI] [Google Scholar]
  • 13.Wang HC, Fussell SR, Cosley D. From Diversity to Creativity: Stimulating Group Brainstorming with Cultural Differences and Conversationally-Retrieved Pictures. In: Proceedings of the ACM 2011 Conference on Computer Supported Cooperative Work. CSCW’11. New York, NY, USA: Association for Computing Machinery; 2011. p. 265–274. Available from: 10.1145/1958824.1958864. [DOI]
  • 14. Bell ST, Villado AJ, Lukasik MA, Belau L, Briggs AL. Getting Specific about Demographic Diversity Variable and Team Performance Relationships: A Meta-Analysis. Journal of Management. 2011;37(3):709–743. doi: 10.1177/0149206310365001 [DOI] [Google Scholar]
  • 15. Uzzi B, Mukherjee S, Stringer M, Jones B. Atypical Combinations and Scientific Impact. Science. 2013;342(6157):468–472. doi: 10.1126/science.1240474 [DOI] [PubMed] [Google Scholar]
  • 16. Lungeanu A, Contractor NS. The Effects of Diversity and Network Ties on Innovations: The Emergence of a New Scientific Field. American Behavioral Scientist. 2015;59(5):548–564. doi: 10.1177/0002764214556804 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Horwitz SK, Horwitz IB. The Effects of Team Diversity on Team Outcomes: A Meta-Analytic Review of Team Demography. Journal of Management. 2007;33(6):987–1015. doi: 10.1177/0149206307308587 [DOI] [Google Scholar]
  • 18. O’Reilly CA III, Williams KY, Barsade S. Group demography and innovation: Does diversity help? In: Composition. Elsevier Science/JAI Press.; 1998. pp. 183–207. [Google Scholar]
  • 19. Tajfel H, Turner JC, Austin WG, Worchel S. An integrative theory of intergroup conflict. Organizational identity: A reader. 1979;56(65):9780203505984–16. [Google Scholar]
  • 20. Harrison DA, Price KH, Bell MP. Beyond Relational Demography: Time and the Effects of Surface- and Deep-Level Diversity on Work Group Cohesion. Academy of Management Journal. 1998;41(1):96–107. doi: 10.2307/256901 [DOI] [Google Scholar]
  • 21. Lau DC, Murnighan JK. Interactions Within Groups and Subgroups: The Effects of Demographic Faultlines. Academy of Management Journal. 2005;48(4):645–659. doi: 10.5465/amj.2005.17843943 [DOI] [Google Scholar]
  • 22. Garrison G, Wakefield RL, Xu X, ‘Kim SH. Globally Distributed Teams: The Effect of Diversity on Trust, Cohesion and Individual Performance. SIGMIS Database. 2010;41(3):27–48. doi: 10.1145/1851175.1851178 [DOI] [Google Scholar]
  • 23. Dobbin F, Kalev A. Why diversity programs fail. Harvard Business Review. 2016;94(7):14. [Google Scholar]
  • 24. Mukherjee S, Huang Y, Neidhardt J, Uzzi B, Contractor N. Prior shared success predicts victory in team competitions. Nature human behaviour. 2019;3(1):74–81. doi: 10.1038/s41562-018-0460-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Salehi N, McCabe A, Valentine M, Bernstein M. Huddler: Convening Stable and Familiar Crowd Teams Despite Unpredictable Availability. In: Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing. CSCW’17. New York, NY, USA: Association for Computing Machinery; 2017. p. 1700–1713. Available from: 10.1145/2998181.2998300. [DOI]
  • 26. Lungeanu A, Huang Y, Contractor NS. Understanding the assembly of interdisciplinary teams and its impact on performance. Journal of Informetrics. 2014;8(1):59–70. doi: 10.1016/j.joi.2013.10.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Gruenfeld DH, Mannix EA, Williams KY, Neale MA. Group Composition and Decision Making: How Member Familiarity and Information Distribution Affect Process and Performance. Organizational Behavior and Human Decision Processes. 1996;67(1):1–15. doi: 10.1006/obhd.1996.0061 [DOI] [Google Scholar]
  • 28. Hinds PJ, Carley KM, Krackhardt D, Wholey D. Choosing Work Group Members: Balancing Similarity, Competence, and Familiarity. Organizational Behavior and Human Decision Processes. 2000;81(2):226–251. doi: 10.1006/obhd.1999.2875 [DOI] [PubMed] [Google Scholar]
  • 29. Huckman RS, Staats BR. Fluid Tasks and Fluid Teams: The Impact of Diversity in Experience and Team Familiarity on Team Performance. Manufacturing & Service Operations Management. 2011;13(3):310–328. doi: 10.1287/msom.1100.0321 [DOI] [Google Scholar]
  • 30. Harrison DA, Klein KJ. What’s the difference? diversity constructs as separation, variety, or disparity in organizations. Academy of Management Review. 2007;32(4):1199–1228. doi: 10.5465/amr.2007.26586096 [DOI] [Google Scholar]
  • 31.Kargar M, An A. Discovering Top-k Teams of Experts with/without a Leader in Social Networks. In: Proceedings of the 20th ACM International Conference on Information and Knowledge Management. CIKM’11. New York, NY, USA: Association for Computing Machinery; 2011. p. 985–994. Available from: 10.1145/2063576.2063718. [DOI]
  • 32. Yu S, Xia F, Zhang C, Wei H, Keogh K, Chen H. Familiarity-Based Collaborative Team Recognition in Academic Social Networks. IEEE Transactions on Computational Social Systems. 2021; p. 1–14. doi: 10.1109/TCSS.2021.3135809 [DOI] [Google Scholar]
  • 33.Contractor N, DeChurch LA, Sawant A, Li X. My Dream Team Assembler; 2013.
  • 34.Knowledge & Data Engineering Group UoK. Benchmark Folksonomy Data from BibSonomy, version of December 31st, 2006; 2006.
  • 35.Gousios G. The GHTorrent dataset and tool suite. In: Proceedings of the 10th Working Conference on Mining Software Repositories. MSR’13. Piscataway, NJ, USA: IEEE Press; 2013. p. 233–236. Available from: http://dl.acm.org/citation.cfm?id=2487085.2487132.
  • 36. Wang X, Zhao Z, Ng W. USTF: A Unified System of Team Formation. IEEE Transactions on Big Data. 2016;2(1):70–84. doi: 10.1109/TBDATA.2016.2546303 [DOI] [Google Scholar]
  • 37. Gómez-Zará D, DeChurch LA, Contractor NS. A Taxonomy of Team-Assembly Systems: Understanding How People Use Technologies to Form Teams. Proc ACM Hum-Comput Interact. 2020;4(CSCW2). [Google Scholar]
  • 38.Valentine MA, Retelny D, To A, Rahmati N, Doshi T, Bernstein MS. Flash Organizations: Crowdsourcing Complex Work by Structuring Crowds As Organizations. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. CHI’17. New York, NY, USA: Association for Computing Machinery; 2017. p. 3523–3537. Available from: 10.1145/3025453.3025811. [DOI]
  • 39. Schildt H. Big data and organizational design—the brave new world of algorithmic management and computer augmented transparency. Innovation. 2017;19(1):23–30. doi: 10.1080/14479338.2016.1252043 [DOI] [Google Scholar]
  • 40. Kellogg KC, Valentine MA, Christin A. Algorithms at Work: The New Contested Terrain of Control. Academy of Management Annals. 2020;14(1):366–410. doi: 10.5465/annals.2018.0174 [DOI] [Google Scholar]
  • 41.Alqahtani M, Gauch S, Salman O, Ibrahim M, Al-Saffar R. Diverse Group Formation Based on Multiple Demographic Features. 2020;.
  • 42. Andrejczuk E, Bistaffa F, Blum C, Rodríguez-Aguilar JA, Sierra C. Synergistic team composition: A computational approach to foster diversity in teams. Knowledge-Based Systems. 2019;182:104799. doi: 10.1016/j.knosys.2019.06.007 [DOI] [Google Scholar]
  • 43. Das A, Gómez-Zará D, Contractor N. Forming Diverse Teams Based on Members’ Social Networks: A Genetic Algorithm Approach. In: Benito RM, Cherifi C, Cherifi H, Moro E, Rocha LM, Sales-Pardo M, editors. Complex Networks & Their Applications IX. Cham: Springer International Publishing; 2021. p. 346–357. [Google Scholar]
  • 44. Layton RA, Loughry ML, Ohland MW, Ricco GD. Design and validation of a web-based system for assigning members to teams using instructor-specified criteria. Advances in Engineering Education. 2010;2(1):n1. [Google Scholar]
  • 45.Zhou S, Valentine M, Bernstein MS. In Search of the Dream Team: Temporally Constrained Multi-Armed Bandits for Identifying Effective Team Structures. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. CHI’18. New York, NY, USA: Association for Computing Machinery; 2018. p. 1–13. Available from: 10.1145/3173574.3173682. [DOI]
  • 46. Ramos-Figueroa O, Quiroz-Castellanos M, Mezura-Montes E, Schütze O. Metaheuristics to solve grouping problems: A review and a case study. Swarm and Evolutionary Computation. 2020;53:100643. doi: 10.1016/j.swevo.2019.100643 [DOI] [Google Scholar]
  • 47. El-Ashmawi WH, Ali AF, Tawhid MA. An improved particle swarm optimization with a new swap operator for team formation problem. Journal of Industrial Engineering International. 2019;15(1):53–71. doi: 10.1007/s40092-018-0282-6 [DOI] [Google Scholar]
  • 48.Bhowmik A, Borkar V, Garg D, Pallan M. Submodularity in team formation problem. In: Proceedings of the 2014 SIAM International Conference on Data Mining. A. B: C; 2014. p. 893–901. Available from: 10.1137/1.9781611973440.102. [DOI]
  • 49. Keane P, Ghaffar F, Malone D. Using machine learning to predict links and improve Steiner tree solutions to team formation problems-a cross company study. vol. 5. Springer; 2020. p. 1–19. [Google Scholar]
  • 50.Agrawal R, Golshan B, Terzi E. Grouping Students in Educational Settings. In: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. KDD’14. New York, NY, USA: Association for Computing Machinery; 2014. p. 1017–1026. Available from: 10.1145/2623330.2623748. [DOI]
  • 51.Nurjanah D, Dewanto K, Sari FD. Homogeneous group formation in collaborative learning using fuzzy C-means. In: 2017 IEEE 6th International Conference on Teaching, Assessment, and Learning for Engineering (TALE); 2017. p. 74–79.
  • 52. Srba I, Bielikova M. Dynamic Group Formation as an Approach to Collaborative Learning Support. IEEE Trans Learn Technol. 2015;8(2):173–186. doi: 10.1109/TLT.2014.2373374 [DOI] [Google Scholar]
  • 53. Agustín-Blas LE, Salcedo-Sanz S, Ortiz-García EG, Portilla-Figueras A, Pérez-Bellido AM, Jiménez-Fernández S. Team formation based on group technology: A hybrid grouping genetic algorithm approach. Computers & Operations Research. 2011;38(2):484–495. doi: 10.1016/j.cor.2010.07.006 [DOI] [Google Scholar]
  • 54. Wi H, Oh S, Mun J, Jung M. A team formation model based on knowledge and collaboration. Expert Systems with Applications. 2009;36(5):9121–9134. doi: 10.1016/j.eswa.2008.12.031 [DOI] [Google Scholar]
  • 55. Zzkarian A, Kusiak A. Forming teams: an analytical approach. IIE transactions. 1999;31(1):85–97. doi: 10.1080/07408179908969808 [DOI] [Google Scholar]
  • 56.Yannibelli VD, Amandi A. Forming well-balanced collaborative learning teams according to the roles of their members: An evolutionary approach. In: 2011 IEEE 12th International Symposium on Computational Intelligence and Informatics (CINTI); 2011. p. 265–270.
  • 57.Lappas T, Liu K, Terzi E. Finding a team of experts in social networks. In: Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining. KDD’09. New York, NY, USA: Association for Computing Machinery; 2009. p. 467–476.
  • 58.Anagnostopoulos A, Becchetti L, Castillo C, Gionis A, Leonardi S. Online team formation in social networks. In: Proceedings of the 21st international conference on World Wide Web. WWW’12. New York, NY, USA: Association for Computing Machinery; 2012. p. 839–848.
  • 59. Meza JLC, Yildirim MB, Masud ASM. A Model for the Multiperiod Multiobjective Power Generation Expansion Problem. IEEE Transactions on Power Systems. 2007;22(2):871–878. doi: 10.1109/TPWRS.2007.895178 [DOI] [Google Scholar]
  • 60.Kargar M, An A, Zihayat M. Efficient bi-objective team formation in social networks. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer; 2012. p. 483–498.
  • 61. Coello Coello CA. Theoretical and numerical constraint-handling techniques used with evolutionary algorithms: a survey of the state of the art. Computer Methods in Applied Mechanics and Engineering. 2002;191(11):1245–1287. doi: 10.1016/S0045-7825(01)00323-1 [DOI] [Google Scholar]
  • 62. Zhou A, Qu BY, Li H, Zhao SZ, Suganthan PN, Zhang Q. Multiobjective evolutionary algorithms: A survey of the state of the art. Swarm and Evolutionary Computation. 2011;1(1):32–49. doi: 10.1016/j.swevo.2011.03.001 [DOI] [Google Scholar]
  • 63. Pérez-Toledano MA, Rodriguez FJ, García-Rubio J, Ibañez SJ. Players’ selection for basketball teams, through Performance Index Rating, using multiobjective evolutionary algorithms. PLOS ONE. 2019;14(9):1–20. doi: 10.1371/journal.pone.0221258 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64. Zhang L, Zhang X. Multi-objective team formation optimization for new product development. Computers & Industrial Engineering. 2013;64(3):804–811. doi: 10.1016/j.cie.2012.12.015 [DOI] [Google Scholar]
  • 65. McPherson M, Smith-Lovin L, Cook JM. Birds of a Feather: Homophily in Social Networks. Annual Review of Sociology. 2001;27(1):415–444. doi: 10.1146/annurev.soc.27.1.415 [DOI] [Google Scholar]
  • 66. Fathian M, Saei-Shahi M, Makui A. A New Optimization Model for Reliable Team Formation Problem Considering Experts’ Collaboration Network. IEEE Transactions on Engineering Management. 2017;64(4):586–593. doi: 10.1109/TEM.2017.2715825 [DOI] [Google Scholar]
  • 67. Monge PR, Contractor NS, Contractor PS, Peter R, Noshir S, et al. Theories of communication networks. Oxford University Press, USA; 2003. [Google Scholar]
  • 68.Eftekhar M, Ronaghi F, Saberi A. Team Formation Dynamics: A Study Using Online Learning Data. In: Proceedings of the 2015 ACM on Conference on Online Social Networks. COSN’15. New York, NY, USA: Association for Computing Machinery; 2015. p. 257–267. Available from: 10.1145/2817946.2817967. [DOI]
  • 69.Horn J, Nafpliotis N, Goldberg DE. A niched Pareto genetic algorithm for multiobjective optimization. In: Proceedings of the First IEEE Conference on Evolutionary Computation. IEEE World Congress on Computational Intelligence; 1994. p. 82–87 vol.1.
  • 70. Gallagher K, Sambridge M. Genetic algorithms: A powerful tool for large-scale nonlinear optimization problems. Computers & Geosciences. 1994;20(7):1229–1236. doi: 10.1016/0098-3004(94)90072-8 [DOI] [Google Scholar]
  • 71. Deb K, Pratap A, Agarwal S, Meyarivan T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Transactions on Evolutionary Computation. 2002;6(2):182–197. doi: 10.1109/4235.996017 [DOI] [Google Scholar]
  • 72.Zihayat M, Kargar M, An A. Two-Phase Pareto Set Discovery for Team Formation in Social Networks. In: 2014 IEEE/WIC/ACM International Joint Conferences on Web Intelligence (WI) and Intelligent Agent Technologies (IAT). vol. 2; 2014. p. 304–311.
  • 73. Paquete L, Chiarandini M, Stützle T. Pareto Local Optimum Sets in the Biobjective Traveling Salesman Problem: An Experimental Study. In: Gandibleux X, Sevaux M, Sörensen K, T’kindt V, editors. Metaheuristics for Multiobjective Optimisation. Berlin, Heidelberg: Springer Berlin Heidelberg; 2004. p. 177–199. [Google Scholar]
  • 74. Dubois-Lacoste J, López-Ibáñez M, Stützle T. Pareto Local Search Algorithms for Anytime Bi-objective Optimization. In: Hao JK, Middendorf M, editors. Evolutionary Computation in Combinatorial Optimization. Berlin, Heidelberg: Springer Berlin Heidelberg; 2012. p. 206–217. [Google Scholar]
  • 75. Zitzler E, Laumanns M, Thiele L. SPEA2: Improving the strength Pareto evolutionary algorithm. TIK-report. 2001;103. [Google Scholar]
  • 76. Zhang X, Guo P, Zhang H, Yao J. Hybrid Particle Swarm Optimization Algorithm for Process Planning. Mathematics. 2020;8(10). doi: 10.3390/math8101745 [DOI] [Google Scholar]
  • 77. Li M, Chen T, Yao X. How to Evaluate Solutions in Pareto-based Search-Based Software Engineering? A Critical Review and Methodological Guidance. IEEE Transactions on Software Engineering. 2020; p. 1–1. doi: 10.1109/TBME.2020.2975279 [DOI] [Google Scholar]
  • 78. Uzzi B, Spiro J. Collaboration and Creativity: The Small World Problem. American Journal of Sociology. 2005;111(2):447–504. doi: 10.1086/432782 [DOI] [Google Scholar]
  • 79. Guimerà R, Uzzi B, Spiro J, Amaral LAN. Team Assembly Mechanisms Determine Collaboration Network Structure and Team Performance. Science. 2005;308(5722):697–702. doi: 10.1126/science.1106340 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Gómez-Zará D, Paras M, Twyman M, Lane JN, DeChurch LA, Contractor NS. Who Would You Like to Work With? In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. CHI’19. New York, NY, USA: Association for Computing Machinery; 2019. p. 1–15. Available from: 10.1145/3290605.3300889. [DOI]
  • 81. Vedres B, Stark D. Structural Folds: Generative Disruption in Overlapping Groups. American Journal of Sociology. 2010;115(4):1150–1190. doi: 10.1086/649497 [DOI] [Google Scholar]
  • 82. Wax A, DeChurch LA, Contractor NS. Self-Organizing Into Winning Teams: Understanding the Mechanisms That Drive Successful Collaborations. Small Group Research. 2017;48(6):665–718. doi: 10.1177/1046496417724209 [DOI] [Google Scholar]
  • 83. Huckman RS, Staats BR, Upton DM. Team Familiarity, Role Experience, and Performance: Evidence from Indian Software Services. Management Science. 2009;55(1):85–100. doi: 10.1287/mnsc.1080.0921 [DOI] [Google Scholar]
  • 84. Staats BR, Gino F, Pisano GP, Edmondson DH, Pierce L, Spektor EM, et al. Varied experience, team familiarity, and learning: The mediating role of psychological safety. Harvard Business School; 2010. [Google Scholar]
  • 85. O’Donovan R, McAuliffe E. Exploring psychological safety in healthcare teams to inform the development of interventions: combining observational, survey and interview data. BMC health services research. 2020;20(1):1–16. doi: 10.1186/s12913-020-05646-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86. Arrow H, McGrath JE, Berdahl JL. Small groups as complex systems: Formation, coordination, development, and adaptation. Sage Publications; 2000. [Google Scholar]
  • 87.Salehi N, Bernstein MS. Hive: Collective Design Through Network Rotation. Proc ACM Hum-Comput Interact. 2018;2(CSCW).
  • 88. Bowers CA, Pharmer JA, Salas E. When Member Homogeneity is Needed in Work Teams: A Meta-Analysis. Small Group Research. 2000;31(3):305–327. doi: 10.1177/104649640003100303 [DOI] [Google Scholar]
  • 89. Mohammed S, Angell LC. Personality Heterogeneity in Teams: Which Differences Make a Difference for Team Performance? Small Group Research. 2003;34(6):651–677. doi: 10.1177/1046496403257228 [DOI] [Google Scholar]
  • 90. Leonardi P, Contractor N. Better people analytics. Harvard Business Review. 2018;96(6):70–81. [Google Scholar]
  • 91.Ye L, Sun H, Wang X, Wang J. Personalized teammate recommendation for crowdsourced software developers. In: Proceedings of the 33rd ACM/IEEE International Conference on Automated Software Engineering. ASE 2018. New York, NY, USA: Association for Computing Machinery; 2018. p. 808–813.
  • 92.Hastings EM, Jahanbakhsh F, Karahalios K, Marinov D, Bailey BP. Structure or Nurture? The Effects of Team-Building Activities and Team Composition on Team Outcomes. Proc ACM Hum-Comput Interact. 2018;2(CSCW).
  • 93.Jahanbakhsh F, Fu WT, Karahalios K, Marinov D, Bailey B. You Want Me to Work with Who? Stakeholder Perceptions of Automated Team Formation in Project-Based Courses. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. CHI’ 17. New York, NY, USA: Association for Computing Machinery; 2017. p. 3201–3212.
  • 94. Chapman KJ, Meuter M, Toy D, Wright L. Can’t We Pick our Own Groups? The Influence of Group Selection Method on Group Dynamics and Outcomes. Journal of Management Education. 2006;30(4):557–569. [Google Scholar]
  • 95. Granovetter MS. The Strength of Weak Ties. American Journal of Sociology. 1973;78(6):1360–1380. doi: 10.1086/225469 [DOI] [Google Scholar]

Decision Letter 0

Sergi Lozano

31 Aug 2021

PONE-D-21-21409

In search of diverse and connected teams: a computational approach to assemble diverse teams based on members' social networks

PLOS ONE

Dear Dr. Gomez-Zara,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

As you can see below, despite being quite positive concerning your manuscript, both reviewers have asked for clarifications regarding a number of aspects. In particular, when revising your text, please pay special attention to the following issues:

- Clarify the contribution of this article beyond that of the conference paper mentioned in the text (as required by Reviewer 1). Notice PLOS ONE's publication criterion #1 (https://journals.plos.org/plosone/s/criteria-for-publication#loc-1), on originality of the presented work.

- Address the concerns by Reviewer 1 on the definition of communication costs in the context of a newly-formed team.

- Justify the methodological aspects highlighted by Reviewer 2 in points 5, 6, 7 and 10. Notice PLOS ONE's publication criterion #3 (https://journals.plos.org/plosone/s/criteria-for-publication#loc-3), on presentation of the methodology.

- Revise interpretations made from the obtained results (especially concerning concerns raised by Reviewer 2 in points 8 and 9 of the report). Notice PLOS ONE's publication criterion #4 (https://journals.plos.org/plosone/s/criteria-for-publication#loc-4), on the link between results and conclusions.

Please submit your revised manuscript by Oct 09 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Sergi Lozano

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please amend your Methods section and Ethics Statement to state whether participants in the DreamTeam data provided informed consent. Please state the type of consent (e.g., written, verbal, etc).

3. Thank you for stating the following in the Acknowledgments Section of your manuscript: 

This work has been supported by the National Science Foundation Awards 560

SES-2021117, SMA-1856090, a 2020 Microsoft Research Dissertation Grant, and 561

National Institutes of Health Awards 1R01GM112938-01, 1R01GM137410-01

Please note that funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. 

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: 

Diego Gomez-Zara (DGZ)

Noshir Contractor (NC)

National Science Foundation Award SES-2021117 (DGZ, NC)

https://www.nsf.gov/awardsearch/showAward?AWD_ID=2021117

National Science Foundation Award SMA-1856090

https://www.nsf.gov/awardsearch/showAward?AWD_ID=1856090&HistoricalAwards=false

2020 Microsoft Research Dissertation Grant (DGZ)

https://www.microsoft.com/en-us/research/blog/2020-microsoft-research-dissertation-grant-supports-students-cutting-edge-work/

National Institutes of Health Awards 1R01GM112938-01 (NC) 

https://reporter.nih.gov/search/BVIUPOLACkaltHszAOD0bg/project-details/8802448

National Institutes of Health Awards 1R01GM137410-01 (NC).

https://reporter.nih.gov/search/1gRS-DiRp06XJhDj5VuMFQ/project-details/9981984

The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Please include your amended statements within your cover letter; we will change the online submission form on your behalf. 

4. We noted in your submission details that a portion of your manuscript may have been presented or published elsewhere. [This is an extended and revised version of a preliminary conference article that was 563 presented in Complex Networks 2020.] Please clarify whether this conference proceeding or publication was peer-reviewed and formally published. If this work was previously peer-reviewed and published, in the cover letter please provide the reason that this work does not constitute dual publication and should be included in the current manuscript.

5. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

6. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well. 

7. Please remove your figures from within your manuscript file, leaving only the individual TIFF/EPS image files, uploaded separately.  These will be automatically included in the reviewers’ PDF.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: N/A

Reviewer #2: I Don't Know

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The paper focuses on the problem of creating teams with a high diversity of member traits and a high degree of prior relationships. The authors adapt a genetic algorithm to the setting of team formation to achieve teams that satisfy the mentioned criteria and evaluate the algorithm on 2 datasets.

The problem is well motivated (although I have some concerns regarding the communication cost which I have described below) and the algorithm is well explained, and I especially appreciate the pseudo-codes.

I ask that the authors further elaborate on why the communication cost in the problem context (a newly formed team) is defined the way it is. I believe communication cost is important and should be taken into account to address the cold start problem. But if a pair of people in a chosen team aren’t connected by an edge, meaning that they do not already immediately know each other, does it matter if they’re 3 hops away from each other or 4? Especially in the context of workplace or student teams, members of a team typically use computer-mediated communication (CMC) tools to communicate with each other. So while the lack of prior immediate connection or experience with working with a teammate can cause communication problems, it would not entail that a person A whose distance to person B is 3 hops, needs to communicate with B through the 2 persons on his/her route to B and thus would have less of a problem than person C who is connected to B through 4 hops (3 persons), as they both would use the same CMC tools.

In the Discussion, references 60 and 61 are cited in support of the hypothesis that teaming up with people with whom one has indirect relationships can potentially promote familiarity and psychological safety but I did not see this point made in those papers.

I ask that the authors comment on how the algorithm can accommodate traits diversifying which is in fact not desirable (e.g., member preferences for team leadership/hierarchy style need to be homogenous within a team).

In their prior conference paper which this manuscript extends, the authors evaluate the algorithm on 2 courses from the myDreamTeam dataset. Are any of the 3 courses discussed in the current manuscript the ones on which they had evaluated their algorithm before? I ask that the authors more clearly indicate which parts in this manuscript have not been published and discussed before.

Reviewer #2: Thank you for the opportunity to review the manuscript entitled "In search of diverse and connected teams: a computational approach to assemble diverse teams based on members' social networks". The authors addressed an important and interesting question of team formation, and proposed a computational approach to assemble teams with high levels of diversity and familiarity at the same time. I find the explanation of the algorithm clearly communicated. I also enjoyed the visualization of team formation with different tools. That being said, I feel this paper can benefit from further clarification. I list my comments and questions below, and hope the authors find them helpful to improve the manuscript.

Theory

1. I like your focus on maximizing skill diversity and minimizing communication costs. But I find your introduction somewhat perplexing. The first paragraph emphasized on forming teams for optimal team performance; the second paragraph talked about efficiency of team building – finding solutions with minimized time and memory (line 44). Then in line 94 you mentioned there are other objectives such as “minimize communication cost, minimize personnel cost, maximize skills present in a team”, and suddenly summarized your aim of maximizing skill diversity and minimizing communication costs. It confused me. Why did you review other objectives of team formation (and how do they relate to your research concern)? Why did you choose to focus on the combination of skill diversity and communication costs rather than other combinations? And how does your objective link to prior focuses on efficiency and outcomes (or other multi-objective algorithms)? Please clarify.

2. Literature review of team diversity. In the introduction you reviewed that diversity is beneficial for team creativity and innovation (line 8-9). Based on this, you encouraged the diversity of individual attributes (e.g., age, gender, race, and skills, line 179) in team formation. I found this questionable. Prior meta-analytical reviews of team diversity have underscored the contingency perspective in the effectiveness of team diversity. Although functional diversity is often found positive, demographic diversity has no or even negative relationship with team outcomes (Bell, Villado, Lukasik, Belau, & Briggs, 2011). Please extend your literature review, and discuss if it makes sense to maximize skill diversity for all sorts of individual attributes in team formation.

3. Theoretical novelty. Can you clarify your contribution and novelty? Is it about the computational approach you used? Yet as you reviewed, this NSGA-II algorithm was used in prior studies already (Pérez-Toledano, et al., 2019). How is your approach different? Or is it about the new objective you proposed of maximizing diversity and minimizing communication cost (line 445)? It’ll help your readers better grasp your contribution.

4. Theoretical implications. I understand that this algorithm can potentially help practitioners assemble teams for this particular purpose. What I am missing here is the implications for team research. Could you explain and elaborate on your contributions to the team literature. For example, how do your findings advance our understanding of team formation such as the formation process?

Data

5. Choice of your datasets. You tested this algorithm on both My Dream Team Builder and bibsonomy (line 372). Whereas I find My Dream Team Builder a highly relevant and unique sample for your research question, I have trouble understanding why the second dataset of bibsonomy is chosen. First, what are the teams in the bibsonomy dataset? Do you count multi-authored publications as teams? More importantly, why does this dataset qualify for testing your algorithm? You proposed this algorithm to build teams with maximized skill diversity and minimized communication costs. But I am not sure if scientific collaborations shared the same objectives. I find it a bit difficult to envision scientists collaborate to maximize the skill diversity of team projects. Also, My Dream Team Builder creates teams instantly. But in scientific collaboration (or teambuilding), authors may join at different stages of the team project. Should this be a concern as well? Please explain why these two datasets were suitable for testing this algorithm.

Analysis and Results

6. Choice of the algorithms for comparison. I wonder why you decided to compare your algorithm with PLS and SPEA-2. I did not see any of these two methods reviewed in the introduction. Instead, in the introduction, you reviewed single-objective algorithms and mentioned other multi-objective algorithms such as the Multi-objective Particle Swarm Optimization (MOPSO) algorithm (Zhang & Zhang, 2013) and the parallel hybrid grouping genetic algorithm (HGGA, Agustín-Blas, et al., 2011). Could you explain why you assessed PLS and SPEA-2, but not the other multi-objective algorithms that you cited? What motivated you to compare these three?

7. In a relevant vein, I also have trouble understanding why you did not compare with those single-objective algorithms that you spent quite some effort to review (line 94). Logically, it also makes sense to see how your multi-objective algorithm outperforms those single-objective counterparts, such as the MCC algorithm you reviewed (line 101) on minimizing communication cost.

8. Evaluation criteria. You concluded that the NSGA-II algorithm outperforms SPEA-2 and PLS (line 419-420). What are the criteria you based on? Did you rely on the visualization of Figure 2 only? Do you use any quantitative results for this conclusion, such as the product of communication cost of diversity index for each algorithm? please add more details here.

9. Interpretation of Figure 2. I find it a little confusing when I read the figures. First, NSGA-II presents many more solutions (the number of dots in the figures). But the variation across different NSGA-II solutions is also very high. Although NSGA-II solutions tend to relatively locate on the top left area, the large variance on both dimensions is concerning. In contrast, the SPEA-2, PLS, or even random options seem much more concentrated. Does this suggest anything about the reliability of NSGA-II results? How can we interpret the dispersion when evaluating its performance? Again, I feel things can be clarified if you can provide a better explanation of the assessment criteria.

10. The test of memory use. Can you explain why it is important to look that the memory and time usage (p. 426-436)? It does not follow your theory part. Does it relate to the objective of using less memory that some prior studies emphasized (line 44)? But you stated repeatedly that the aims of team building should be maximized skill diversity and minimized communication costs. Also, the difference in memory use (ranging from 1.2 MB to 2.7 MB) does not seem very significant to me. Does it make a big difference in reality? Please clarify why the test is relevant and why this difference is valuable.

REFERENCE

Bell, S. T., Villado, A. J., Lukasik, M. A., Belau, L., & Briggs, A. L. (2011). Getting specific about demographic diversity variable and team performance relationships: A meta-analysis. Journal of management, 37(3), 709-743.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Yingjie Yuan

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Nov 9;17(11):e0276061. doi: 10.1371/journal.pone.0276061.r002

Author response to Decision Letter 0


11 Mar 2022

February 28, 2022

Dear editor and reviewers,

Thank you for providing us with a thorough and constructive review. We found the points identified to be very helpful in revising and improving the paper. We appreciate that the reviewers saw the merits of this manuscript. We are pleased to submit our revised paper, which has been rewritten to address the concerns raised in the initial review.

R1 raised concerns about the definition of communication costs. In this revision, we provide a more elaborated explanation of why we used this metric and why it is appropriate in the team formation problem. We also ran the team formation implementation using alternative metrics proposed by Lappas et al. (2009), which are in the Supplementary Information. In addition, we calculated the number of hops within team members of the simulated teams to assess the distance among members. We found that most team members are connected directly or through two hops. High-level hops should not be a concern. We included this analysis in the Supplementary Information. In the Discussion section, we now indicate how this team formation problem can accommodate traits that require homogeneity rather than heterogeneity. We also fixed the citations, re-established the MDT case presented in the conference paper, and indicated the extensions and non-published parts of this version compared to the conference paper.

R2 requested several clarifications and justifications across the paper’s sections. We re-wrote the Introduction section to provide more clarity and state the paper’s contributions. Based on prior studies, we elaborated on the conditions and settings when team diversity can be beneficial. Moreover, we describe prior studies that indicate the benefits of demographic diversity and functional diversity. We acknowledge that team diversity itself is not enough to enhance teams’ outcomes, so we describe the purpose of considering team familiarity in this problem. Then, we establish our team formation problem as theoretically novel since it considers more than one objective function and assigns all individuals into teams. Our paper differs from prior “best-team” and single-objective formulations. We also elaborate on the theoretical implications of this work for team science. In particular, the role of technologies in the team formation process.

We also justified the choices of datasets and algorithms used in this paper in more detail. In addition to these new justifications, we added a new algorithm based on particle swarm optimization that aligns with the literature review and a different dataset that includes software developer teams formed on GitHub. Regarding the work that we reviewed, we explain why some of these algorithms and problem formulations are incompatible with our proposed optimization problem. While most of these algorithms consider either a single-objective function or forming one single team, our proposal considers multiple objective functions and assigns all available individuals into teams.

Furthermore, we included quantitative metrics to evaluate the final Pareto front results, which have been highly used in prior studies. We also explain why the variety of solutions (rather than the convergence in a specific set of solutions) is important for multi-objective problems. Finally, we replaced the test of memory with a test of time complexity.

In the remainder of this document, we address each of the points raised by the reviewers and explain the steps we have taken to address these concerns in the revised manuscript.

We appreciate the opportunity to revise and resubmit our paper, and we sincerely hope we can share our contribution at PLOS One. Thank you.

Sincerely,

The Authors

Editor’s meta-review

Clarify the contribution of this article beyond that of the conference paper mentioned in the text (as required by Reviewer 1). Notice PLOS ONE's publication criterion #1, on the originality of the presented work.

In this revised version, we clarify the contributions and extensions of this article beyond the conference paper (Page 4). We also state the theoretical and practical implications of this work in the Introduction and Discussion sections. We confirm that the extended material of this paper has not been published in any previous publications. More details are provided in response to R1 in this letter (Page 5).

Address the concerns by Reviewer 1 on the definition of communication costs in the context of a newly-formed team.

We have addressed R1’s concerns on the definition of communication costs and its interpretation in the context of team formation. The clarification is in this response letter on page 3 and in the manuscript on page 8.

Justify the methodological aspects highlighted by Reviewer 2 in points 5, 6, 7, and 10. Notice PLOS ONE's publication criterion #3, on the presentation of the methodology.

In this revised version, we have updated and clarified the choices of datasets (Point 5), the choices of algorithms (Point 6), the use of multi-objective functions (Point 7), and the metrics used (Point 10). Beyond the clarifications provided in this new revised version, we:

Added a new dataset to test our problem implementation (Point 5).

Added a new algorithm based on a hybrid version of particle swarm optimization (HPSO) to test our proposed algorithm (Point 6).

Added the hypervolume and the unique non-dominated front ratio metrics to provide a quantitative evaluation of the algorithms (Point 10).

We also confirm that we have revised our methodological aspects guided by PLOS ONE’s publication criteria #3. We describe methods and reagents in sufficient detail for another researcher to reproduce the experiments described.

Revise interpretations made from the obtained results (especially concerning concerns raised by Reviewer 2 in points 8 and 9 of the report). Notice PLOS ONE's publication criterion #4, on the link between results and conclusions.

We revised and updated our interpretations made from the obtained results. In particular, we now include quantitative metrics (i.e., Hypervolume, Nondominated Front Size, coverage) to support our results (Point 8) and explain why the variation among results is important for the algorithms (Point 9). We provide more details and justifications on the response to R2 on page 8. We also confirm that we have revised the link between our results and conclusions according to PLOS ONE's publication criterion #4. We also avoid overstating our conclusions.

Reviewer 1

I ask that the authors further elaborate on why the communication cost in the problem context (a newly formed team) is defined the way it is. I believe communication cost is important and should be taken into account to address the cold start problem. But if a pair of people in a chosen team aren’t connected by an edge, meaning that they do not already immediately know each other, does it matter if they’re 3 hops away from each other or 4? Especially in the context of workplace or student teams, members of a team typically use computer-mediated communication (CMC) tools to communicate with each other. So while the lack of prior immediate connection or experience with working with a teammate can cause communication problems, it would not entail that a person A whose distance to person B is 3 hops, needs to communicate with B through the 2 persons on his/her route to B and thus would have less of a problem than person C who is connected to B through 4 hops (3 persons), as they both would use the same CMC tools.

We use communication cost as a proxy for familiarity among team members: two nodes are connected by an edge if the members have communicated or collaborated before (Kargar et al., 2013). We followed this definition since it has been employed in the team formation algorithms literature to determine collaborations among individuals (Farhadi et al., 2012; Kargar & An, 2011; Lappas et al., 2009). This definition focus on the importance of collaborations and familiarity between experts by taking into account the cost of their collaborations (which Lappas et al. (2009) referred to as communication). According to their model, fewer hops represent lower costs of collaboration. The lowest value possible is when all team members know each other (i.e., they are directly connected), and the highest one is when all members are not connected at all.

The goal of using the sum of distances as a communication cost metric is to operationalize the number of direct connections and shared connections within a team. While one-hops between team members mean that they already worked together in the past, two-hops mean that members had common prior collaborations in the past. This approach is grounded on triadic closure, which posits that nodes are likely to establish a new connection when they have a connection in common. Three-hops and 4-hops can follow the same principles based on balance theories. For example, if user A has worked with user B, user B has worked with user C, and user C worked with user D, then it is likely for user A (or user B) to establish a new connection with user D. Therefore, using communication cost definition allowed our objective function to search for teams that maximized the number of direct collaborations (i.e. no hops), common connections (one hop), and close connections (two hops and more).

Regarding the context of CMC tools, Lappa et al.’s communication cost definition differs from routing problems. This definition does not require that individuals communicate with each other only through their hops, it only refers to how their prior collaboration networks were constituted. In other words, if user A has a 3-hop distance to user B, it means that they are connected by two people who worked together in the past. It does not mean that user A and user B will communicate through their contacts while they work as a team. This definition focuses on their collaborations rather than their communication channels.

To understand the effect of three and higher levels hops, we now report the social network structures contained in these databases. We did not provide this information in the initial submission and we added it to this revised submission to provide more context about people’s relationships (Table 2). Overall, individuals were closely related to each other, and the average diameter of these networks (i.e. the longest short path among all members) was 5.3 for the MDT databases, 3.66 for the bibsonomy databases, and 3.0 for the GHTorrent databases.

We also calculated the frequency of direct contacts, 1-hop, 2-hops, 3-hops in the assembled teams to understand the distances among team members (See S1 Table). As result, we found that the vast majority of the members were connected with others with 2 hops (~31%), followed by direct contacts (~30%). These numbers show that these members were highly connected in general, and a high number of hops was not commonly seen. We hope this information can clarify the minimum effect produced by large hops on the team formation problem.

We acknowledge that there are many other possible ways to define communication costs among members. One alternative is using weighted graphs in which edges represent how familiar team members are with each other: a low-weight edge implies that members have collaborated a few times, while a high-weight edge implies that members have collaborated multiple times. Other alternatives for communication cost are the diameter (i.e., the largest shortest path between any two nodes in the graph), and the minimum spanning tree (i.e., the minimum sum of the weights of a graph’s edges) (Lappas et al., 2009). We repeated our objective function using these definitions and the results are consistent with the ones reported in the main paper. We have included these alternative definitions in the Supplementary Materials.

These clarifications, analyses, and observations are now included in our Methods, Results, and Discussion sections.

In the Discussion, references 60 and 61 are cited in support of the hypothesis that teaming up with people with whom one has indirect relationships can potentially promote familiarity and psychological safety but I did not see this point made in those papers.

Thank you for pointing out this issue. We realized that we made a mistake when we added the citations. The original reference was: “Huckman, R. S., Staats, B. R., & Upton, D. M. (2009). Team familiarity, role experience, and performance: Evidence from Indian software services. Management Science, 55(1), 85-100.”, and the reference in the initial submission was “Huberman, B. A., & Hogg, T. (1995). Communities of practice: Performance and evolution. Computational & Mathematical Organization Theory, 1(1), 73-92.” There was an error when we typed “Hu” in the bibliography manager.

Reference #61 was also not appropriate for supporting the hypothesis that teaming up with familiar members can promote psychological safety. We have addressed this assertion by citing these two papers:

Staats, B. R., Gino, F., Pisano, G. P., Edmondson, D. H., Pierce, L., & Spektor, E. M. (2010). Varied experience, team familiarity, and learning: The mediating role of psychological safety.

These authors ran an experimental study in which participants must resolve multiple tasks and demonstrated that “ team familiarity would positively influence psychological safety” (Page 15).

O’Donovan, R., & McAuliffe, E. (2020). Exploring psychological safety in healthcare teams to inform the development of interventions: combining observational, survey and interview data. BMC health services research, 20(1), 1-16.

In their results, the authors found that “familiarity between team members facilitated psychological safety. Team members found it easier to speak openly as they got to know one another better and worked together for longer.” (Page 8).

We have fixed these citations and edited our assertions in our revised manuscript.

I ask that the authors comment on how the algorithm can accommodate traits diversifying which is in fact not desirable (e.g., member preferences for team leadership/hierarchy style need to be homogenous within a team).

The current algorithm’s implementation supports more than two objective functions. Given the current implementation, we can add a third objective function that minimizes the variance among the team members’ attributes to be considered (e.g., minimize leadership style’s coefficient of variation). This objective function to homogenize members’ attributes can be defined as minimizing the coefficient of variation for continuous variables, or the Blau index for categorical variables. Then, the NSGA-II algorithm will find Pareto fronts that consider (a) lower communication costs, (b) higher diversity in a specific attribute set, and (c) lower diversity in a second attribute set. We added this potential extension in the Discussion section (Page 24).

In their prior conference paper which this manuscript extends, the authors evaluate the algorithm on 2 courses from the myDreamTeam dataset. Are any of the 3 courses discussed in the current manuscript the ones on which they had evaluated their algorithm before?

Only one course discussed in the current manuscript was evaluated in our prior conference paper. We changed the courses in the current manuscript to test different course levels (undergraduate, graduate, and MBA courses). We put back the missing MDT course from the conference paper to our revised paper. Since this implementation now considers individuals without prior connections (i.e., isolates), the plots and numbers presented in this version are not the same ones presented in our conference paper.

I ask that the authors more clearly indicate which parts in this manuscript have not been published and discussed before.

The parts from this manuscript that have not been published and discussed before are:

The datasets’ generation code and the algorithms’ source code.

The pre-processed and de-identified datasets.

The algorithms’ literature review. We describe the team formation literature in-depth to provide a better motivation and elaborated context for this problem. We include previous algorithms that find one single team and team combinations and algorithms.

The algorithm’s pseudo-code and explanation of each component step by step. We revised the code and its parts to improve the readability of the code. We provide more detailed descriptions. In contrast, the conference paper provided only the pseudo-code for the crossover step.

This revised version includes an implementation in which :

(a) the number of nodes can be different from a multiple of the team size, leaving a smaller team with the remaining participants; and

(b) considers members who do not have any edges in the social graph, these members are only considered in the diversity objective function and excluded from the communication cost objective function.

Evaluation with the bibsonomy dataset and GHTorrent dataset. We included these two datasets to prove that our algorithm can work in other team formation domains.

Comparison of our implementation against other benchmark algorithms frequently cited in the literature: PLS, SPEA-2, and HPSO.

We compare the results based on quantitative metrics (hypervolume and the unique non-dominated front ratio) running time used by this algorithm compared to other Pareto-front implementations (PLS, SPEA-2, and HPSO).

We discuss how this implementation can benefit team builders (e.g., managers, instructors) and the consequences of using this team formation algorithm in real teams.

We also enumerate and make explicit these differences in our revised Introduction section.

Reviewer 2

1. I like your focus on maximizing skill diversity and minimizing communication costs. But I find your introduction somewhat perplexing. The first paragraph emphasized on forming teams for optimal team performance; the second paragraph talked about efficiency of team building – finding solutions with minimized time and memory (line 44). Then in line 94 you mentioned there are other objectives such as “minimize communication cost, minimize personnel cost, maximize skills present in a team”, and suddenly summarized your aim of maximizing skill diversity and minimizing communication costs. It confused me. Why did you review other objectives of team formation (and how do they relate to your research concern)? Why did you choose to focus on the combination of skill diversity and communication costs rather than other combinations? And how does your objective link to prior focuses on efficiency and outcomes (or other multi-objective algorithms)? Please clarify.

We appreciate this observation. We rewrote the Introduction section to provide more clarity and focus on maximizing diversity and familiarity. We removed the mentions of team optimal performance, efficiency in team building, and other objective functions from the Introduction. And we created a new section for our team formation algorithms review.

We chose these two objective functions because both diversity and familiarity can be determined during the team formation process, and they both determine team composition. Research shows that the interaction of diversity and familiarity can positively influence team performance. While other approaches (e.g., coaching, training, leadership) can help diverse teams work better, those require interventions after teams are assembled. Therefore, we examine how maximizing diversity and familiarity simultaneously can leverage team formation processes and teams’ composition. We elaborated on these reasons in the Introduction section.

Finally, we acknowledge that making strong statements about the “good” or “bad” effects of diversity in teams is a flawed approach (Bell et al., 2011). In this revised version, we explained better when diversity can be beneficial for teams’ efficiency and outcomes, and removed overstatements of the effects of diversity on team performance. In particular, we point out that prior research has shown the benefits of diversity for creativity and innovation. We link our objective goals on the interaction effect between familiarity and diversity on performance (Huckman et al., 2009). These changes are reflected in the new Introduction section.

2. Literature review of team diversity. In the introduction you reviewed that diversity is beneficial for team creativity and innovation (line 8-9). Based on this, you encouraged the diversity of individual attributes (e.g., age, gender, race, and skills, line 179) in team formation. I found this questionable. Prior meta-analytical reviews of team diversity have underscored the contingency perspective in the effectiveness of team diversity. Although functional diversity is often found positive, demographic diversity has no or even negative relationship with team outcomes (Bell, Villado, Lukasik, Belau, & Briggs, 2011). Please extend your literature review, and discuss if it makes sense to maximize skill diversity for all sorts of individual attributes in team formation.

In line 179, we mentioned examples of individuals’ attributes that could be included in the team formation problem. Our formulation emphasizes the use of categorical and numerical variables that can be chosen by the administrator of the algorithm. These attributes can be in the surface-level (e.g., age, race, gender) and deep-level (e.g., backgrounds, careers, functions, expertise) (Harrison et al., 1998). Although we mentioned these attributes as examples, we did not aim to encourage the use of any particular attributes. From our evaluations, the MyDreamTeam dataset was the only one that included demographic variables. We did not include demographic variables in the bibsonomy simulations. And in our third dataset, we only included skills variables. We clarify in the Discussion section that administrators or managers who will administer this algorithm should reflect and decide on adding demographic variables and functional characteristics according to their organizational goals and particular context (Page 25).

Although prior research on the effects of demographic diversity on team performance has shown mixed results (as we stated in Line 524), several studies have demonstrated their benefits. For example, one study found a positive relationship between gender diversity and team productivity in software engineer teams (Vasilescu et al., 2015), another study found that multi-cultural teams are more likely to provide more creative solutions than teams from a single culture (H.-C. Wang et al., 2011), collective intelligence studies have demonstrated a link between the number of female members and performance (Woolley et al., 2015), and a final example shows that racially diverse teams can compete better than homogeneous teams (Andrevski et al., 2014). Although Bell et al. 2011 found either no relationship between demographic diversity and team performance or small effects, they provide possible explanations and interpretations of these results: (a) how prior studies operationalized race and gender diversity as a variety metric (i.e., number of members of a specific race/sex in the team) rather than separation (i.e., to what extend members seem themselves as a team or not); (b) the context of the study; (c) and the exacerbation of in-group/out-group biases, stereotypes, and prejudices among team members. Overall, there is a consensus that other factors and processes (e.g., familiarity, leadership, perceived diversity, psychological safety, cohesion) moderate the effect of demographic diversity on team performance. Andreviski et al. (2004) found that racial diversity only had a positive effect on team performance when team members had a low aversion against someone from a different race. For these reasons, we consider familiarity as a second objective for team formation because prior relationships in a team can positively moderate the effect of diversity on team performance (Huckman et al., 2009). We added these studies, the advantages and disadvantages of diversity, and the possible effects of moderator variables in the Introduction section.

We also acknowledge that diversity can be beneficial for certain types of tasks only (e.g., creativity tasks, ideation tasks, making-a-decision tasks), which require the combination of different points of view, backgrounds, and experiences (McGrath, 1984). For this reason, we explicitly mention in the Introduction section the benefits of diversity for tasks that require creativity and innovation. We also mention in the Discussion section that the algorithm can also consider optimizing homogenous attributes (e.g., specialization, personality styles) by adding another objective function.

Lastly, we acknowledge in the Limitations subsection that this approach requires experimentation with real teams to test whether this team formation problem leverages teams’ performance. We have elaborated on these restrictions, scopes, and caveats in the Discussion section.

3. Theoretical novelty. Can you clarify your contribution and novelty? Is it about the computational approach you used? Yet as you reviewed, this NSGA-II algorithm was used in prior studies already (Pérez-Toledano, et al., 2019). How is your approach different? Or is it about the new objective you proposed of maximizing diversity and minimizing communication cost (line 445)? It’ll help your readers better grasp your contribution.

The main contribution of this paper is the formulation of the team formation problem considering teams' diversity levels and members' familiarity simultaneously. As a result, team builders can explore different team combination alternatives, and examine the trade-off between familiarity and diversity. While most studies in team formation algorithms have considered members' skills or personal costs as team formation objective functions (X. Wang et al., 2016), we formulate this optimization problem based on different operationalizations of diversity (i.e., disparity and variety of attributes). This formulation allows choosing various and multiple diversity factors that fit organizational goals (e.g., functional, educational, gender). The second contribution of this work is the design of algorithms for this team formation problem that assigns all available individuals to a team. Previous team formation problems have mainly focused on finding the best team from a pool of individuals and dismissing the rest (Gómez-Zará et al., 2020; X. Wang et al., 2016). This “best-team” approach could not fit organizational goals that require all individuals to belong to a group (e.g., workshops, training classes, location assignment).

We outlined these contributions in the Introduction and Discussion sections.

4. Theoretical implications. I understand that this algorithm can potentially help practitioners assemble teams for this particular purpose. What I am missing here is the implications for team research. Could you explain and elaborate on your contributions to the team literature. For example, how do your findings advance our understanding of team formation such as the formation process?

The main theoretical implication of this work is the use of computational mechanisms to support team formation processes. Literature has characterized team formation centered on behavioral mechanisms, where teams can be assembled by internal or external forces and based on similarity, familiarity, and competence (Arrow et al., 2000; Hinds et al., 2000). This optimization problem and the implemented algorithms found diverse and connected team combinations that otherwise individuals could not have otherwise foreseen. This work allows team scholars to reflect on the role of technologies in enabling new organizational structures among individuals and organizations, which can lead to new theories of team formation and technologies (Kellogg & Valentine, 2020; Schildt, 2017; Valentine et al., 2017). We elaborate on this implication in the Introduction and Discussion sections.

5. Choice of your datasets. You tested this algorithm on both My Dream Team Builder and bibsonomy (line 372). Whereas I find My Dream Team Builder a highly relevant and unique sample for your research question, I have trouble understanding why the second dataset of bibsonomy is chosen. First, what are the teams in the bibsonomy dataset? Do you count multi-authored publications as teams? More importantly, why does this dataset qualify for testing your algorithm? You proposed this algorithm to build teams with maximized skill diversity and minimized communication costs. But I am not sure if scientific collaborations shared the same objectives. I find it a bit difficult to envision scientists collaborate to maximize the skill diversity of team projects. Also, My Dream Team Builder creates teams instantly. But in scientific collaboration (or teambuilding), authors may join at different stages of the team project. Should this be a concern as well? Please explain why these two datasets were suitable for testing this algorithm.

The purpose of using bibsonomy was only for evaluation purposes. We used this dataset to test the algorithms’ results and running time. A literature review on team formation algorithms shows DBLP, Bibsonomy, IMDB, and GitHub as appropriate examples for evaluation purposes (H.-C. Wang et al., 2011). We chose bibsonomy since some team formation papers tested their algorithms using this database (Anagnostopoulos et al., 2010, 2012). These papers used co-authorship networks as a proxy of relationships and the paper’s topics as authors’ skills. In our example, two authors are connected if they co-authored at least one paper. The skills were calculated based on the papers’ topics. Given that the topics were very broad, we selected the 20 most frequent topics in each journal and computed authors’ skills based on those 20 topics.

In this revision, we include a third database of GitHub repositories provided by GHTorrent. We perform the same exercise assuming that users can create teams based on repositories. Like the previous exercise, the purpose of using this dataset is to demonstrate the algorithms’ capabilities and results.

We acknowledge that forming scientific teams and software teams is complex in reality: new members can be added over time, some specialization is required, not all of them share the same objectives and restrictions, and diversity may be beneficial for certain goals. We believe these restrictions should not be a concern because we use them only to test the algorithms’ efficiency and results. In the revised version of this paper, we acknowledge the restrictions and limitations of these datasets in the Discussion section.

6. Choice of the algorithms for comparison. I wonder why you decided to compare your algorithm with PLS and SPEA-2. I did not see any of these two methods reviewed in the introduction. Instead, in the introduction, you reviewed single-objective algorithms and mentioned other multi-objective algorithms such as the Multi-objective Particle Swarm Optimization (MOPSO) algorithm (Zhang & Zhang, 2013) and the parallel hybrid grouping genetic algorithm (HGGA, Agustín-Blas, et al., 2011). Could you explain why you assessed PLS and SPEA-2, but not the other multi-objective algorithms that you cited? What motivated you to compare these three?

We recognize that we explained neither the PLS and SPEA-2 algorithms in detail nor why we used them. In this new revision, we add their description and the justification of using them in the Methods section (Pages 16 and 17). Multiple studies have used PLS and SPEA-2 to evaluate and compare multi-objective algorithms (Pérez-Toledano et al., 2019; Zhou et al., 2011; Zihayat et al., 2014). For that reason, we used these algorithms to test our NSGA-II implementation.

Although we mentioned MOPSO and HGGA to provide examples of algorithms for team formation problems, these algorithms cannot be implemented in our diversity & familiarity team formation problem. Zhang & Zhang’s MOPSO implementation only forms one single team (rather than multiple teams) and it finds the solution in a continuous space. Each solution represents whether a member i belongs to the best team or not. Solutions move in a two-dimensional space, and they apply a sigmoid function to binarize the final outcome. In contrast, our team formation problem represents a combinational problem. Our goal is to assign every available individual to a team and test different team combinations. We operationalize team membership using chromosomes. Therefore, their implementation cannot be used for our particular team formation problem.

We checked for an alternative MOPSO solution for combinational problems. We found and developed Zhang et al.’s HPSO implementation (2020), which is a hybrid implementation of MOPSO. HPSO combines the particle swarm optimization steps with evolutionary approaches. We decided to use this alternative MOPSO implementation as another benchmark. We explain how this algorithm works in the Methods section and elaborate on its results in the Discussion section.

Lastly, HGGA cannot be implemented for this particular problem since it was designed for a single-objective problem.

We acknowledge that our literature review did not follow a clear rationale. In this revised version, we restructured the Related Work to emphasize the focus on previous team formation algorithms and why they cannot be used for our particular optimization problem.

7. In a relevant vein, I also have trouble understanding why you did not compare with those single-objective algorithms that you spent quite some effort to review (line 94). Logically, it also makes sense to see how your multi-objective algorithm outperforms those single-objective counterparts, such as the MCC algorithm you reviewed (line 101) on minimizing communication cost.

The problem with single-objective algorithms is that they only provide one single team. Since they do not implement dominance criteria, the single-objective algorithm will prioritize the best team on one dimension given specific restrictions or constraints. Our goal with this multi-objective implementation is to help team builders examine team combinations with different trade-offs.

Regarding the MCC algorithm, it aims to find the best team possible given a pool of individuals. It will only provide a team of size n, dismissing the other individuals. In contrast, our approach teams up all the individuals of the pool. We could implement their specific problem using an evolutionary algorithm but it would only provide one single solution. If that is the case, we expect that an MCC implementation will find a combination located in one of the extremes of the Pareto Front if the trade-off parameter is 0 or 1 (i.e., highest diversity, lowest communication cost), or in the middle of the Pareto Front if the tradeoff parameter is 0.5. We provide this clarification in the Related Work section (Pages 6 and 7).

8. Evaluation criteria. You concluded that the NSGA-II algorithm outperforms SPEA-2 and PLS (line 419-420). What are the criteria you based on? Did you rely on the visualization of Figure 2 only? Do you use any quantitative results for this conclusion, such as the product of communication cost of diversity index for each algorithm? please add more details here.

Thank you for pointing out this. We relied on the visualization of Figure 2 and the time/resources tests in the original version. In this revised version, we include quantitative metrics (i.e., Hypervolume, Nondominated Front Size, coverage) to support our results. The metrics are explained in the Methods section (Page 18), and their results are discussed in the Results section (Pages 19-21).

9. Interpretation of Figure 2. I find it a little confusing when I read the figures. First, NSGA-II presents many more solutions (the number of dots in the figures). But the variation across different NSGA-II solutions is also very high. Although NSGA-II solutions tend to relatively locate on the top left area, the large variance on both dimensions is concerning. In contrast, the SPEA-2, PLS, or even random options seem much more concentrated. Does this suggest anything about the reliability of NSGA-II results? How can we interpret the dispersion when evaluating its performance? Again, I feel things can be clarified if you can provide a better explanation of the assessment criteria.

In this revision, we provide a better explanation of the assessment criteria and the importance of diverse solutions (Cao et al., 2015). Pareto Fronts’ shapes provide useful information about the amount of tradeoff between different dimensions (e.g., communication cost, diversity), and how much compromise is needed from some criteria to improve others. Finding the true Pareto front of this team formation problem is computationally hard given we need to compute and assess all the possible team combinations. For this reason, algorithms use a series of steps to find an approximation of the true Pareto front. A critical assumption for these algorithms is that the Pareto Front is sufficiently populated. The quality of this approximation depends upon (1) the proximity of the points on the approximated front to the points on the true Pareto front; and (2) the diversity of the solutions on the approximated front, where more diversity is typically better. Although the true Pareto front is unknown, solutions that dominate others are close to the theoretical true Pareto Front. Then, the diversity of the solutions will provide a larger range and granularity of the Pareto Front.

That said, the large variance on both dimensions shows that the NSGA-II algorithm found more non-dominated solutions, which is desirable and not concerning. The crowding distance step of NSGA-II allowed the algorithm to keep a broader range of non-dominant solutions. Additionally, the algorithm kept secondary solutions in different layers that could have originated non-dominant solutions in later iterations. As the algorithm continues creating new generations, non-dominant solutions can be still considered to find other potential solutions. Moreover, NSGA-II can still identify non-dominant solutions in the middle of the trade-off.

In contrast, the low variance of the other algorithms shows that they converged on a specific set of solutions and to a specific trade-off. They do not consider other possible combinations that prioritize either familiarity or diversity. Therefore, these algorithms can lack diverse solutions that reside in the extremes of both dimensions.

We have articulated a better explanation of a Pareto front in the Related Work section (Page 6) and the interpretations of the results in the Result section.

10. The test of memory use. Can you explain why it is important to look that the memory and time usage (p. 426-436)? It does not follow your theory part. Does it relate to the objective of using less memory that some prior studies emphasized (line 44)? But you stated repeatedly that the aims of team building should be maximized skill diversity and minimized communication costs. Also, the difference in memory use (ranging from 1.2 MB to 2.7 MB) does not seem very significant to me. Does it make a big difference in reality? Please clarify why the test is relevant and why this difference is valuable.

We recognize that these metrics were not theoretically driven and elaborated without enough details. In this new version, we removed the memory analysis and explain the importance of checking algorithms’ running-time based on computational complexity.

As the results show, the running times of PLS and HPSO increase exponentially as the input size increases. In contrast, NSGA-II provides solutions faster than the other algorithms, and NSGA-II’s time increase stays at a quadratic scale (O(n^2)). Moreover, NSGA-II required less than one-third of the time that PLS and HPSO took to provide similar results. Therefore, using NSGA-II is highly encouraged to find solutions efficiently as the input size increases. These differences should be considered in addition to the Pareto Front metrics discussed in the revised manuscript. We included these clarifications and interpretations in our Results and Discussion sections.

References

Anagnostopoulos, A., Becchetti, L., Castillo, C., Gionis, A., & Leonardi, S. (2010). Power in unity: forming teams in large-scale community systems. ACM. https://doi.org/10.1145/1871437.1871515

Anagnostopoulos, A., Becchetti, L., Castillo, C., Gionis, A., & Leonardi, S. (2012). Online team formation in social networks. Proceedings of the 21st International Conference on World Wide Web, 839–848. https://doi.org/10.1145/2187836.2187950

Andrevski, G., Richard, O. C., Shaw, J. D., & Ferrier, W. J. (2014). Racial Diversity and Firm Performance: The Mediating Role of Competitive Intensity. Journal of Management, 40(3), 820–844. https://doi.org/10.1177/0149206311424318

Arrow, H., McGrath, J. E., & Berdahl, J. L. (2000). Small groups as complex systems: Formation, coordination, development, and adaptation. Sage Publications.

Backstrom, L., Boldi, P., Rosa, M., Ugander, J., & Vigna, S. (2012). Four degrees of separation. Proceedings of the 4th Annual ACM Web Science Conference, 33–42. https://doi.org/10.1145/2380718.2380723

Bell, S. T., Villado, A. J., Lukasik, M. A., Belau, L., & Briggs, A. L. (2011). Getting Specific about Demographic Diversity Variable and Team Performance Relationships: A Meta-Analysis. Journal of Management, 37(3), 709–743. https://doi.org/10.1177/0149206310365001

Cao, Y., Smucker, B. J., & Robinson, T. J. (2015). On using the hypervolume indicator to compare Pareto fronts: Applications to multi-criteria optimal experimental design. Journal of Statistical Planning and Inference, 160, 60–74. https://doi.org/10.1016/j.jspi.2014.12.004

Contractor, N. (2001). Small worlds: The dynamics of networks between order and randomness. Chance, 14(4), 33–36.

Farhadi, F., Hoseini, E., Hashemi, S., & Hamzeh, A. (2012). TeamFinder: A Co-clustering based Framework for Finding an Effective Team of Experts in Social Networks. 2012 IEEE 12th International Conference on Data Mining Workshops, 107–114. https://doi.org/10.1109/ICDMW.2012.54

Gómez-Zará, D., DeChurch, L. A., & Contractor, N. S. (2020). A Taxonomy of Team-Assembly Systems: Understanding How People Use Technologies to Form Teams. Proceedings of the ACM on Human-Computer Interaction, 2(CSCW2), 36. https://doi.org/10.1145/3415252

Harrison, D. A., Price, K. H., & Bell, M. P. (1998). Beyond relational demography: Time and the effects of surface-and deep-level diversity on work group cohesion. Academy of Management Journal. Academy of Management. https://journals.aom.org/doi/abs/10.5465/256901

Hinds, P. J., Carley, K. M., Krackhardt, D., & Wholey, D. (2000). Choosing work group members: Balancing similarity, competence, and familiarity. Organizational Behavior and Human Decision Processes, 81(2), 226–251.

Huckman, R. S., Staats, B. R., & Upton, D. M. (2009). Team Familiarity, Role Experience, and Performance: Evidence from Indian Software Services. Management Science, 55(1), 85–100. https://doi.org/10.1287/mnsc.1080.0921

Kargar, M., & An, A. (2011). TeamExp: Top-k Team Formation in Social Networks. 2011 IEEE 11th International Conference on Data Mining Workshops, 1231–1234. https://doi.org/10.1109/ICDMW.2011.162

Kargar, M., Zihayat, M., & An, A. (2013). Finding Affordable and Collaborative Teams from a Network of Experts. In Proceedings of the 2013 SIAM International Conference on Data Mining (SDM) (pp. 587–595). Society for Industrial and Applied Mathematics. https://doi.org/10.1137/1.9781611972832.65

Kellogg, K. C., & Valentine, M. A. (2020). Algorithms at work: The new contested terrain of control. Academy of Management. https://journals.aom.org/doi/abs/10.5465/annals.2018.0174

Lappas, T., Liu, K., & Terzi, E. (2009). Finding a team of experts in social networks. Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 467–476. https://doi.org/10.1145/1557019.1557074

Milgram, S. (1967). The small world problem. Psychology Today, 2(1), 60–67. http://files.diario-de-bordo-redes-conecti.webnode.com/200000013-211982212c/AN%20EXPERIMENTAL%20STUDY%20by%20Travers%20and%20Milgram.pdf

Pérez-Toledano, M. Á., Rodriguez, F. J., García-Rubio, J., & Ibañez, S. J. (2019). Players’ selection for basketball teams, through Performance Index Rating, using multiobjective evolutionary algorithms. PloS One, 14(9), e0221258. https://doi.org/10.1371/journal.pone.0221258

Schildt, H. (2017). Big data and organizational design – the brave new world of algorithmic management and computer augmented transparency. Innovation (North Sydney, N.S.W.), 19(1), 23–30. https://doi.org/10.1080/14479338.2016.1252043

Uzzi, B., Amaral, L. A. N., & Reed-Tsochas, F. (2007). Small-world networks and management science research: a review. European Management Review, 4(2), 77–91. https://doi.org/10.1057/palgrave.emr.1500078

Valentine, M. A., Retelny, D., To, A., Rahmati, N., Doshi, T., & Bernstein, M. S. (2017). Flash Organizations: Crowdsourcing Complex Work by Structuring Crowds As Organizations. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (pp. 3523–3537). Association for Computing Machinery. https://doi.org/10.1145/3025453.3025811

Vasilescu, B., Posnett, D., Ray, B., van den Brand, M. G. J., Serebrenik, A., Devanbu, P., & Filkov, V. (2015). Gender and Tenure Diversity in GitHub Teams. 3789–3798. https://doi.org/10.1145/2702123.2702549

Wang, H.-C., Fussell, S. R., & Cosley, D. (2011). From Diversity to Creativity: Stimulating Group Brainstorming with Cultural Differences and Conversationally-retrieved Pictures. Proceedings of the ACM 2011 Conference on Computer Supported Cooperative Work, 265–274. https://doi.org/10.1145/1958824.1958864

Wang, X., Zhao, Z., & Ng, W. (2016). USTF: A Unified System of Team Formation. IEEE Transactions on Big Data, 2(1), 70–84. https://doi.org/10.1109/TBDATA.2016.2546303

Woolley, A. W., Aggarwal, I., & Malone, T. W. (2015). Collective Intelligence and Group Performance. Current Directions in Psychological Science, 24(6), 420–424. https://doi.org/10.1177/0963721415599543

Zhang, X., Guo, P., Zhang, H., & Yao, J. (2020). Hybrid Particle Swarm Optimization Algorithm for Process Planning. Science in China, Series A: Mathematics, 8(10), 1745. https://doi.org/10.3390/math8101745

Zhou, A., Qu, B.-Y., Li, H., Zhao, S.-Z., Suganthan, P. N., & Zhang, Q. (2011). Multiobjective evolutionary algorithms: A survey of the state of the art. Swarm and Evolutionary Computation, 1(1), 32–49. https://doi.org/10.1016/j.swevo.2011.03.001

Zihayat, M., Kargar, M., & An, A. (2014). Two-Phase Pareto Set Discovery for Team Formation in Social Networks. 2014 IEEE/WIC/ACM International Joint Conferences on Web Intelligence (WI) and Intelligent Agent Technologies (IAT), 2, 304–311. https://doi.org/10.1109/WI-IAT.2014.112

Attachment

Submitted filename: Response Letter.pdf

Decision Letter 1

Sergi Lozano

15 Jul 2022

PONE-D-21-21409R1In search of diverse and connected teams: a computational approach to assemble diverse teams based on members' social networksPLOS ONE

Dear Dr. Gomez-Zara,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. As you can see below, both reviewers have recommended publication. However, Reviewer #1 has asked for some intuition on the quality metrics introduced, and has pointed out some minor issues. Please, address them in your revised version. Moreover, since PLOS ONE does not proof-read manuscript, I suggest revising the text for small typos and shortcomings. For instance, the word 'otherwise' is written twice in line 979.

Please submit your revised manuscript by Aug 29 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Sergi Lozano

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: N/A

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you for responding to the reviews. I find the approach, the choice of the algorithm, and the communication cost generally better motivated compared to the initial submission. The new Related Work section is now also better connected to the rest of the paper.

I appreciate the authors clarifying the points that I had asked for including the definition of communication cost as well as clarifying the average network diameter and frequency of direct contacts.

One request I have is for the authors is to provide intuition about the metrics that they have used for evaluating the NSGA-II algorithm and why they matter in the specific context of team formation. For example, intuitively, why is a higher hypervolume value preferred and what qualities about a selection of teams does it demonstrate?

Minor points:

Line 318-319 is awkwardly phrased and I had a hard time reading it at first. The clause “and it has been classified as an NP-hard problem” in effect refers to the minimization problem but with the sentence it is conjoined with, it seems as if it refers to computing the communication cost which in fact, can be done in polynomial time.

Line 296: P_j & P_j => P_i & P_j

Reviewer #3: The authors have approached an interesting topic in a novel way, pairing group dynamics with computer science. The methods and analyses are sound, and the results are intriguing. I would support publication of this manuscript in its current form.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Nov 9;17(11):e0276061. doi: 10.1371/journal.pone.0276061.r004

Author response to Decision Letter 1


8 Sep 2022

September 9th, 2022

Dear Dr. Lozano and reviewers,

We thank you and appreciate your precious time and efforts in revising our manuscript. Your valuable and insightful observations allowed us to improve this article. We are pleased to submit our revised paper to address the most recent comments.

R1 requested providing more intuition about the metrics used to evaluate the algorithms. We added more context and rationale in the Metrics sub-section and Results section. Also, we have fixed the typos identified by R1. In addition, two authors have proofread this new version and fixed typos and grammar errors.

We also want to thank R3 for the kind comments and for taking the time to revise this resubmission.

In the remainder of this document, we address the points raised in the last review and explain the steps we have taken to address these concerns in this revised manuscript.

We appreciate the opportunity to revise and resubmit our paper, and we sincerely hope we can share our contribution at PLOS One. Thank you.

Sincerely,

The Authors

---------

Editor’s meta-review

As you can see below, both reviewers have recommended publication. However, Reviewer #1 has asked for some intuition on the quality metrics introduced, and has pointed out some minor issues. Please, address them in your revised version.

>> In this revised version, we elaborate on the rationale of the metrics used in this study. More details are provided in response to R1 in this letter.

Moreover, since PLOS ONE does not proof-read manuscript, I suggest revising the text for small typos and shortcomings. For instance, the word 'otherwise' is written twice in line 979.

>> Two authors have proofread this new version and fixed typos and grammar errors.

-----------------

Reviewer 1

One request I have is for the authors is to provide intuition about the metrics that they have used for evaluating the NSGA-II algorithm and why they matter in the specific context of team formation. For example, intuitively, why is a higher hypervolume value preferred and what qualities about a selection of teams does it demonstrate?

>> In this revised version, we provide more insight into the metrics used by this study. In summary, higher hypervolume scores show that teams with higher levels of diversity and familiarity can be found. The Unique Non-dominated Front Ratio shows how many high-diversity/high-familiarity team combinations were found by only one algorithm. Lastly, evaluating the running time is essential since some problems will require large participant pools to be assigned to teams. We develop the rationale and intuition of these metrics in the Metrics sub-section and Results section.

Line 318-319 is awkwardly phrased and I had a hard time reading it at first. The clause “and it has been classified as an NP-hard problem” in effect refers to the minimization problem but with the sentence it is conjoined with, it seems as if it refers to computing the communication cost which in fact, can be done in polynomial time.

>> Thank you for pointing out this error. It was in fact referring to the minimization problem and not to computing the communication cost. We removed the sentence and the resulting paragraph is the following: “The goal is to minimize the average sum of shortest path lengths across all assembled teams in the individuals’ network. Computing the sum of communication costs of a set of teams runs in O(n2) time.”

Line 296: P_j & P_j => P_i & P_j

>> Fixed.

Attachment

Submitted filename: Response to Reviewers.pdf

Decision Letter 2

Seyedali Mirjalili

28 Sep 2022

In search of diverse and connected teams: a computational approach to assemble diverse teams based on members' social networks

PONE-D-21-21409R2

Dear Dr. Gomez-Zara,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Seyedali Mirjalili

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: N/A

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you for clarifying the intuition behind the metrics used for evaluating the algorithm, and for making edits throughout the paper to ensure better readability. I am glad to recommend the paper for acceptance.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Supporting figures and tables.

    S1 Fig: Simulations using the Diameter metric. S2 Fig: Simulations using the Minimum Spanning Tree (MST) metric. S1 Table: Diameter Case. S2 Table: Minimum Spanning Tree Case. S3 Table: Team combinations’ average proportion of hops.

    (PDF)

    Attachment

    Submitted filename: Response Letter.pdf

    Attachment

    Submitted filename: Response to Reviewers.pdf

    Data Availability Statement

    We have deposited a minimal data set in the GitHub repository https://nusoniclab.github.io/. This repository contains the following files: (1) the pre-processed and de-identified data used in this study, (2) the Python scripts to pre-process the original datasets, (3) the Python scripts that run the proposed algorithm and benchmark algorithms, including their plots and quantitative metrics. The pre-processed data in this repository constitutes the minimal dataset required to replicate all study findings reported in this article. The first dataset (MyDreamTeam) is administered by the SONIC Research Group, Northwestern University. We deposited a de-identified and minimal dataset from MyDreamTeam in the Github Repository to replicate the findings reported in this article. Because of the sensitive nature of some of the variables collected, Northwestern University Institutional Review Board (IRB)-approved protocol does not permit individual-level data to be made unrestricted and publicly available. Researchers interested in obtaining a larger sample of this restricted and de-identified individual-level data should contact the authors to inquire about receiving an IRB-approved institutional data sharing agreement. The second dataset used (bibsonomy) is administered by the Knowledge and Data Engineering Group, University of Kassel. This dataset is available under a license agreement, and it can be requested at https://www.kde.cs.uni-kassel.de/wp-content/uploads/bibsonomy/. While we do not maintain this dataset, we provide the scripts to generate the pre-processed datasets used in this study. The third dataset (GHTorrent) is administered by Georgios Gousios. The dataset is freely and publicly available at https://ghtorrent.org/. While we do not maintain these datasets, we provide the scripts to generate the pre-processed datasets used in this study.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES