Skip to main content
PLOS One logoLink to PLOS One
. 2021 Jul 14;16(7):e0250301. doi: 10.1371/journal.pone.0250301

Analyzing the relationship between productivity and human communication in an organizational setting

Arindam Dutta 1,*, Elena Steiner 1, Jeffrey Proulx 2, Visar Berisha 1, Daniel W Bliss 1, Scott Poole 2, Steven Corman 1
Editor: Nersisson Ruban3
PMCID: PMC8279403  PMID: 34260597

Abstract

Though it is often taken as a truism that communication contributes to organizational productivity, there are surprisingly few empirical studies documenting a relationship between observable interaction and productivity. This is because comprehensive, direct observation of communication in organizational settings is notoriously difficult. In this paper, we report a method for extracting network and speech characteristics data from audio recordings of participants talking with each other in real time. We use this method to analyze communication and productivity data from seventy-nine employees working within a software engineering organization who had their speech recorded during working hours for a period of approximately 3 years. From the speech data, we infer when any two individuals are talking to each other and use this information to construct a communication graph for the organization for each week. We use the spectral and temporal characteristics of the produced speech and the structure of the resultant communication graphs to predict the productivity of the group, as measured by the number of lines of code produced. The results indicate that the most important speech and network features for predicting productivity include those that measure the number of unique people interacting within the organization, the frequency of interactions, and the topology of the communication network.

Introduction

The “structural imperative” in network research [1] suggests that we can represent any organization as a network and look at the network as a determinant of behavior, culture, and the individuals within the organization. Organizational networks are generated and populated by human beings who are active agents with intentions, knowledge, and the ability to rationalize their actions. From interactions between individuals in an organization we can derive certain qualitative aspects like behavior, intentions, emotions and inter-employee relations of a workplace. These aspects play a large role in the effectiveness and productivity of an organization. In this paper we aim to directly study this relationship between productivity and communication, and report new methods for doing so.

While productivity is relatively straightforward to measure, existing studies measure communication indirectly, either through member self-reports of communication on rating scales [2, 3], through external raters’ evaluation using global scales that assess communication behavior [3], as communication technology investment [4], or through questionnaires measuring more distal constructs such as communication satisfaction or perceived effectiveness [5, 6]. While these studies are useful, they can be challenged on the grounds that perceptions of communication do not correspond to actual communication behavior [7]. Direct observation is the “gold standard” for measuring communication and provides the most rigorous test of the communication-productivity relationship. Though several studies involving direct observation of communication behavior have been completed (for a review see [8]), these typically involved methods of human observation of small groups for short periods or unusual settings (for example Ham radio operators) where communication is routinely logged. Long-term studies based on objective observation are needed to supplement and validate current understanding of the relationship between communication and productivity.

Our general research question is:

What is the relationship between the amount of communication in an organization and its productivity? What are the factors that may moderate this relationship?

Several factors may moderate the productivity-communication relationship. One particularly important factor is the type of work the organizational unit in question does. For units engaged in the production of verbal outputs-such as plans, reports, audits and in those whose primary work involves interacting with clients or customers-such as those delivering education, therapy, or advice-an argument can be made that the greater the amount of communication, the higher the productivity. For units engaged in action or production, however, a different relationship would be expected: communication is good up to a point, but too much communication interferes with action or production. Moreover, in these units, high levels of communication may signal that they are experiencing difficulties and hence must engage in problem solving that requires high levels of communication. In this case, we can expect a non-linear relationship between communication and productivity, communication is positively related to productivity up to a point, past which it is negatively related. Since the organizational unit we are studying is engaged in producing software, we would expect an inverted-U shaped (2nd order polynomial) relationship between communication and productivity.

In this work, we estimate inter-employee communication networks in a software engineering organization using speech recordings. For a period of 3 years, all employees wore audio-recorders during their hours of work which recorded their conversations, and weekly communication graphs were estimated based on the detected speech. We use a simple speech activity detector, combined with inter-recorder correlations, to detect interactions between individuals and to construct daily communication graphs. In addition, we also measure several speech features that describe the speaking style of each individual. These features, which are defined in more detail in the S1 Appendix, include, pitch, temporal features (energy, zero crossing rate), spectral features (spectral centroid, spectral flux etc), and cepstral features (mel-scale frequency cepstral coefficients-MFCCs). Numerous studies have used these speech features to detect speakers and speech features such as emotions with high accuracy [917]. Each research has, in turn, linked various speech features to emotion. At the neurological level, emotions are known to have an impact on individual task performance [18, 19]. Emotion also influences individual behavior in task performance, citizenship and deviance [20]. Ashforth and Humphrey [21] reviewed the importance of emotion in organizational contexts, including its effects on motivation, leadership, and group dynamics. All of these have been associated with performance in empirical research, for example, motivation, [22], leadership [23] and group dynamics [24]. It is important to study emotion alongside network structure because networks are a substrate of emotional contagion, and such contagion has been shown to influence group dynamics [25]. Therefore, we use a combination of networks and speech analysis to analyze the relationship between productivity and human communication in an organization. The method for this study was not intended to be applied by other organizations for practical purposes. Our immediate purpose in comparing productivity to detected interaction was to validate our detection method, i.e. to prove that the communication we detected has expected relationships to organizational outcomes. An additional purpose was to support a larger sponsored project, focused on discrepancies between observable and perceived communication [26].

Method

Organization setting and data collection

This study was approved by the Arizona State University IRB (Approval number: STUDY00003138), and written consent forms were obtained for participants. The setting for this research was the Software Factory (SF), a service unit at a large southwestern university providing software engineering services for funded research projects and university technology spinouts. SF had directors and work was led by a professional software engineer who managed student programmers using industry-standard engineering processes and were organized in forma, project-based teams. These characteristics put it squarely in the category of a professional organization [27]. It operated for 144 weeks from late 2002 to early 2005, and had 79 participants, including the manager, employees, clients, and researchers. Over this time, SF worked on 31 separate projects, developing applications for the social sciences, natural sciences, and education, and for internal use (such as an activity reporting system). The major steps of handling a project at the Software Factory consisted of four major processes:

  • The business process,

  • The development process,

  • The design process, and

  • The implementation process.

Typically, the initial business process involved the most senior people on the customer side (including the decision maker) and the highest-level SF personnel (one or more directors and a project manager). When the client had already identified one or more students to work on the project, they may also be in attendance. The development process included collaboration between the customer, project manager and the technical lead of the project. The major activities in this process involved validating with the customer, setting realistic customer expectations, and communicating to all SF personnel working on the project. The design process included the project manager, technical lead and the developers, and lastly the implementation stage involved the technical lead and the developers. These projects varied in terms of timescale and the number of SF personnel involved. Over the course of 144 weeks, there were instances where multiple projects existing at the same time, involving multiple employees, and some instances with an employee being involved in multiple projects at the same time. This study used only records from the 54 SF employees, because only employees made entries in a code repository and activity reporting system, data used in this paper.

The SF data is a unique dataset that aimed to accomplish, as nearly as possible, ubiquitous observation of a set of 79 employees and clients of the organization. The dataset contains recorded audio data from participants between September-2002 and June-2005. Whenever they entered the dedicated SF facility, participants attached a digital recorder and lapel microphone, and logged in to a server which placed a time stamp on the recording. When leaving, they uploaded the recorded audio to a server for storage. The resultant dataset contains daily recordings of all SF employees and visitors (primarily clients) comprising approximately 7000 hours of time synchronized recordings. There was no evidence if employees ever chose to delete or not turn in recordings, it would have been reflected in our time-aligning analyses for cross-correlation mentioned in the later section. Also, people involved in SF said that after the first week or so, members tended to forget the recorders. The same has been reported in other studies doing long-term recording of participants. The participant recordings were created in digital speech standard (DSS) file formats, a compressed proprietary format optimized for speech. They were converted to an uncompressed WAV format using the Switch Sound File Converter software. The files were stored using a 6kHz sampling rate with 8-bits/sample.

In addition to the recordings, we analyzed the code written by employees at the SF. All codes were stored and managed using a Visual Source Safe (VSS) 6.0 repository. We used the VSS API to extract records from the repository. Each record included the filename, date, user, version, and changes, insertions, and deletions at check-in. From this information we were able to compute the number of lines of code at each check-in. In particular, we computed the total number of inserted, deleted and changed lines of code per employee per week. A total of 11276 entries of changes in LOC were recorded staring from the first week of March-2003.

The SF dataset affords a unique opportunity to obtain a holistic picture of work activity and communication in a small organizational unit over an extended period. In this analysis, we have used the audio recording from March-2003 to June-2005 (124 weeks), to build communication networks and extract speech features to predict the effective lines of codes obtained using VSS analysis.

Other studies in the literature have found that LOC is an effective measure of productivity in software organizations [28, 29].

Approach

All analyses were done on a weekly basis. In case of communication graphs, individual interactions between any two individuals were detected using a simple cross-correlation scheme. Individual interactions were converted to a communication graph representing the frequency of interactions between any two individuals over the course of a week. From this graph, we extracted a set of features that describe the topology of the resultant network and denote that by, GwR1×fg, where fg is total number of graph features. In addition, we also extracted several speech features from the daily recordings and calculate two statistics (mean and variance) for these features across the whole week for all participants. These are defined as, SwR1×(2×fs), where fs is total number of speech features. Thus, we had a total communication feature space defined by Cw:(GwSw)R1×(fg+2×fs) (where ‖ is the concatenation operator).

We describe the details of how we estimate the communication graph and the feature extraction in the sections below. We then describe how we predict productivity using these features.

Communication graph analysis

Pair-wise communication detection

To construct the communication graphs, we used cross channel signal analysis. The entire process of graph analysis can be subdivided into two main blocks, the construction of speech cross-correlation graphs and graph feature extraction as shown is Fig 1.

Fig 1. Process chain for communication graph analysis.

Fig 1

Speech cross correlation graph

As a pre-processing step we normalized the data by the mean to remove DC offset (caused by the analogue parts of the system that add a DC current to the audio signal), that causes significant interference with the audio signal, especially during signal processing. We investigated preliminary conversation detection performance on the SF data by using a two-stage approach. The first stage identified continuous segments of speech using an energy and spectral based detector; in the second stage, we use a pair-wise cross-correlation between one speaker’s channel and the remaining channels to detect with whom that person was speaking. The basic idea behind this approach is that, if two individuals are speaking, their microphone will pick up each-other’s speech and cross correlation will be high. A cross-correlation matrix was constructed using mean correlation weights between participant pairs across each day. The weights were calculated based the quantity of communication between participant pairs for an entire working day. The correlation matrix represents a proxy for the frequency of interactions between any two individuals. The same data can also be used to detect individual interactions and compare against manually coded data. Pairwise conversations between two speakers were detected by the algorithm and were presented to research assistants for manual coding. The daily cross-correlation matrices, which represent a proxy for frequency of interaction between two speakers, were averaged over the week to construct weighted communication graphs, with participants as nodes and the correlation weights as edges.

In the automated interaction detection, we used simple speech processing techniques from audio segments of both employees in a dyad to detect communication. First, we computed the short-time speech energy and spectral centroid (See S1 Appendix) for every 15 seconds frame and estimated thresholds to detect speech from the two features. Speech portions were detected using the two thresholds and non-speech portions were removed.

Next, we computed the covariance matrix between energy of speech segments from both microphones in a dyad. Two sets of thresholds were estimated based on the diagonal elements of the matrix, (a) Th1, to determine if communication occurred (0 or 1, 2, 3) and (b) Th2, to determine the direction of communication (1, 2 or 3).

Validation of detection

Before constructing the communication graphs based on pair-wise cross-correlation, we validated the detections by comparing them to human coder classifications of the audio recordings as indicating network connections. We extracted 10 minute audio segments from a dyad from random working days. First we determined the total number of segments required to assess validity. Based on this we extracted that number of segments through random sampling from the audio corpus. External raters then coded the 15 second segments regarding whether there was talk or silence in the segment and who was talking to whom. The specific classifications they could make were:

  • Silence/noise (0)

  • Employee 1 speaking (1)

  • Employee 2 speaking (2)

  • Both employees speaking (3)

We determined the minimum number of audio segments required to assess validity using the confidence interval equation,

Np^(1-p^)ϵ2

where N is the minimum number of samples, p^ is the estimated population proportion and ϵ is the margin of error. With an error margin (variance per sample) of 5% and a p^ of 0.8, the minimum number of samples required is 64. In our analyses, a total of 75 ten minutes audio segments from random working days and between random dyads were used for communication validation. As Fig 3 indicates, there was 88% agreement between the coders and the automated detection (see next section for more details).

Graph feature extraction

After the graph was constructed using pairwise speech correlation, we extracted several topological features that aim to describe the nature of daily interactions. A total of 11 graph features were investigated in this work, which are described in more details in S1 Appendix.

Basic graph descriptors. We calculated the following basic graph descriptors:

  • Number of edges. The total number of communication links present between employees in the network.

  • Number of nodes. The total number of active employees present in the network.

  • Average degree. Defined as the number of links that are incident on a particular employee. It is informative of total communication for individual employees.

  • Number of connected triples. A count of the number of connected triples in the graph.

  • Number of cycles in a graph. Defined as mn + c, where m is the number of links, n is the number of employees and c is the number of connected components. This indicates how connected the network is.

  • Graph energy. The sum of the absolute values of the real components of the eigenvalues of the graph. They tell us about the structural complexity of the network. A structurally complex network has more differentiated interactions, which suggests members are working on different tasks in smaller groups and also that there is some interchange among these small groups.

Graph centrality measures. We computed the following graph centrality measures:

  • Degrees. The average number of links adjacent to an employee node. This is an effective measure of the influence or importance of individual nodes on the network.

  • Average neighbor degree. The average degree of adjacent or neighboring nodes for every vertex. We took the average of this measure across all nodes. This indicates the flow of communication around the organizational unit.

  • Eigen centrality. The i-th component of the eigenvector of the adjacency matrix gives the centrality score of the i-th node of the network. The average eigen centrality across all nodes was computed for this study. This measure tells us about the quality of communication of an employee with others. This indicates the influence an employee over other employees in the organization.

Laplacian features. We also calculated two Laplacian graph features.

  • Graph spectrum. Defined as the eigenvalues of the Laplacian of the graph. This tells us about the frequency of communication in the organizational unit and its relationship to the nodes and link attributes.

  • Algebraic connectivity. The magnitude of this value reflects how well connected the overall network is. It has been used in analyzing the robustness and synchronizability of networks.

These features are estimated based on daily graphs. We average over the week to compute a weekly graph feature vector, GwR1×11, where 11 is total number of graph features investigated.

Speech analysis

In addition to the graph features, we extracted speech features for every speaker from the data. These features carry information about speaker identity and various aspects of affect, which are important characteristics for predicting productivity.

Speech feature extraction

Speech features are extracted independently for every speaker (e.g. every recording channel). Prior to feature extraction, we remove the DC offset, and split the data into 1-second speech segments using hamming windows. All features are extracted at this scale.

A total of 35 different features were obtained from the audio data. Some of these pertained to whether there was a network linkage between actors and others pertained to properties of the linkages. In view of the exploratory nature of this research, we included the latter in order to capture a richer description of the nature of the links than a simple linked-not linked description would provide. As mentioned before, emotion affects productivity and these emotions can be recognized from variations in various aspects of speech. The speech features used for this study are mentioned below and described in details in S1 Appendix,

  • Pitch. Features related to pitch contain information related to speaker emotions [9, 10, 13]. Fundamental pitch frequency, 12 harmonics and harmonic ratio were the pitch-related features that were investigated in this study.

  • Temporal features. These features capture certain aspects of speaker emotion, like stress level, joy, excitement etc [9, 10]. We calculated the zero-crossing rate, shot-time energy and energy entropy from every one-second speech frame.

  • Spectral features. These features carry the particulars of the frequency content of speech. They carry information about speaker identity and can help classifying a wide range of emotions [10, 11]. The spectral features investigated in this study are the spectral centroid, spectral spread, spectral entropy, spectral flux and spectral rolloff.

  • Cepstral features. These features capture the characteristics of our auditory system based on changes in emotions, irrespective of language or gender. A significant number of speech emotion recognition (SER) research papers have identified these as one of the most efficient features for emotion classification [911, 13, 16]. Thirteen Mel-frequency cepstrum coefficients (MFCC) were extracted from 20 ms frames and averaged over 1 sec window.

We calculated the mean and variance of these features over the working days of a week to compute weekly speech feature vectors defined as, SwR1×(35×2), where 35 is total number of graph features investigated and 2 is the number of statistics computed for each speech feature. Thus, together with the graph and speech features we had a combined communication feature set defined by, CwR1×(11+70)=R1×81.

Measure of productivity

In this paper, the overall organization productivity, defined by the total lines of codes per week per employee (LOCw) was used as the measure of productivity in the SF. The total LOC was calculated for each week as the sum of ‘changed’, ‘inserted’ and ‘deleted’ LOC, as, LOCw = Changed + inserted + deleted LOC. The weekly LOC measures were converted to log scale to reduce the variable dynamic range. The average LOC per employee was calculated bu normalizing the LOC measure by the number of employees present during the particular week.

Predicting productivity from communication

Regression methods allow us to summarize and study relationships between two continuous (quantitative) variables. One variable is regarded as the predictor, explanatory, or independent variable (in this case the weekly ‘communication features, Cw’), and the other variable, is regarded as the response, outcome, or dependent variable (in this case weekly ‘productivity, LOCw’). We mentioned before that we should expect an inverted U-shaped relationship (polynomial of order 2) between communication and productivity. To apply this hypothesis, we first selected the communication features that exhibited such relationship. The selected communication features were then used to predict the organizational productivity. Since the variables are consecutive and evenly-spaced observations in time, it is a sequence of discrete-time data, where each data point is dependent on previously observed values. Consequently, We used a time-series regression model to predict productivity. In general, our regression model assumes productivity and the communication features are related to one another by

LOCw=F(Cw,t)+ϵ

where F(Cw,t) is some mathematical operation (or model) showing productivity as a function of the input communication features and time (weeks), and ϵ is the prediction error. Fig 2 shows the block diagram of the prediction process and each block is described.

Fig 2. Prediction process chain.

Fig 2

Pre-whitening

Pre-whitening is required to remove autocorrelation and trends from the time-series variables, so that a meaningful relationship between the variables can be assessed. It concentrates the main variance in the data in a relatively small number of dimensions, and removes all first-order structure from the data. We implemented the ZCA whitening transformation,

X^=X-μ(X)cov(X)

where, μ(X) and cov(X) are the mean and covariance matrix of time-series variable X. X^ is the transformed variable whose covariance matrix is the identity matrix. We pre-whitened all the independent variables (Cw) and the dependent variable (LOCw).

Feature selection

We used a rank based feature selection method with a regression model (F(Cw)) to evaluate correlation weights of each communication feature independently with 10-fold cross validation (in a 10-fold cross validation, the entire set is divided into 10 subsets, where 9 of them are used to train the regression model and one set for prediction). A support vector regression (SVR) model (see S1 Appendix) with a second order polynomial kernel (according to hypothesis) was used to find the association of each feature with the measure of productivity. Features with correlation weights above zero were selected for prediction analysis. Fig 2 shows that 27 communication features were selected from 81, which were given as input to the regression model.

Time-series regression

After selecting the most correlated features, they were used to predict productivity (LOCw) using a time-series regression model. The SVR model with second order polynomial kernel was used as the base regression model. We can write the final model as

F(Cw,t)=kF(Cw(t-k))

To test the accuracy of the model k-steps ahead predictions were made at each data point, for k = 0, 1, 2, 3, …, 8. Prediction for various time lags (1–8 weeks) were evaluated, to assess the dependency of productivity on past data.

Results

Pair-wise communication detection results

In the pair-wise communication detection, the four main classes were, “Silence/noise” (0), “Employee 1 speaking” (1), “Employee 2 speaking” (2), and “Both employees speaking” (3). The receiver operating characteristics (ROC) curve (see Fig 3) was used to illustrate the communication detection accuracy (0 or 1, 2, 3). The ROC curve was constructed by varying the threshold Th1, and the optimum value of Th1 was determined. Threshold Th2 was determined after constructing confusion matrices for various Th2 values. The threshold parameters for the best model were Th1 = 2.53e−5 and Th2 = 2.02e−5. We have shown the confusion matrix of the best detection model in Table 1.

Fig 3. Receiver operating characteristics curve for communication detection; area under curve (AUC) = 0.88.

Fig 3

Table 1. Confusion matrix for the best detection model; each element is shown in terms of number of 15 seconds segments.

Coder
Class 0 1 2 3
Tool 0 1390 183 227 74
1 51 201 32 105
2 70 14 309 98
3 16 42 50 138

Our method produced a good communication detection rate (AUC: 0.88), and on reviewing the results, we noticed that most of the false positives resulted because of the presence of other employees. We then constructed the daily communication graph using the above detection method, with correlation weights as edges connecting the employees present in the day. Thus in case of a communication scenario with more than two employees, the correlation weights will be high for any dyad with the speaker in it, while the correlation weights between other employees will be relatively low. For any focal individual the correlation weights between that individual will be high with anyone they address, while those between other speakers who might be detected in the background is lower.

Predicting productivity from communication

Feature selection

We computed the correlation weights for each communication feature while predicting productivity. Fig 4 shows the average merit of the features based on correlation weights achieved while predicting LOCw. It can be seen that almost all the graph features (10 out of 11) had positive correlation weights. Among the weekly speech features, the MFCC coefficients (1, 2, 3, 4, 5, 6, 8), the spectral and energy entropy (mean), fundamental frequency (variance), spectral roll-off (mean) and spectral centroid and spread (mean) were positively correlated. Comparing the two types of communication features, the graph features had higher correlation weight than the speech features. The number of nodes, average neighbor degree, algebraic connectivity, graph energy and graph spectrum were the features with highest average merit.

Fig 4. Correlation coefficients for each selected communication feature.

Fig 4

Time-series prediction of productivity

To analyze the communication-productivity relationship we made k-steps time-series prediction of LOCw at each data point using the selected communication features. We used lags of upto six weeks to analyze how much the productivity depend on previous weeks’ communication. The mean absolute error (MAE), mean absolute percentage error (MAPE), root mean squared error (RMSE) and direction accuracy (DA) were measured to evaluate the accuracy of the time-series model. The time-series model implementation was done in WEKA 3.8 [30]. Fig 5 shows the k-steps (k = 1, 2, 4, 8) prediction result using a lag of one week. The accuracy parameters are shown in Table 2 for 1 week and 6 weeks lags. Fig 6 shows the MAPE for different lags (1 to 8 weeks).

Fig 5. 1, 2, 4, 8 steps ahead prediction for productivity (LOCw) using 1-week past information.

Fig 5

Table 2. Accuracy of time-series model used to predict productivity (LOCw) using communication features (Cw).
MAE RMSE MAPE % DA %
1 week lag 1-step 0.64 1.36 7.2 77.5
2-steps 0.83 9.18 9.2 74.5
4-steps 0.96 1.80 9.9 71.3
8-steps 0.97 1.79 9.8 73.1
6 weeks lag 1-step 0.17 0.58 2.2 92.5
2-steps 0.25 0.75 2.8 89.5
4-steps 0.34 0.83 3.2 89.4
8-steps 0.51 1.07 5.9 83.8

MAE: Mean absoulte error; RMSE: Root mean square error; MAPE: Mean absolute percentage error; DA: Direction accuracy.

Fig 6. Mean absolute percentage error while predicting productivity (LOCw) for different lags (1–8 weeks).

Fig 6

It can be seen that, using 1-week previous information, we can predict productivity (LOCw) with an error of 7.2–9.8% (1–8 steps ahead prediction). This is error is reduced to 2.2–5.6%, when we use information from the previous 6 weeks. The direction accuracy also improves from 71–77% to 83–92%.

Discussion

From the results we can conclude that communication is strongly related to productivity in an organization. Table 2 suggests that we can predict organizational productivity with high accuracy with mean absolute error less than 10%. We hypothesized before, that communication and productivity share a non-linear relationship (polynomial of order 2), and we made use of that relationship in the regression model. With the use of a second order polynomial kernel SVR model, we selected the communication features and used to same model to do a time-series forecasting of productivity. The results are also suggestive of the fact that the prediction accuracy improved as we used more previous information. Though comparisons are difficult due to differences in methods and measures, this study shows a stronger correlation between communication and performance than previous research. In [6], the authors found a relationship of r = 0.27 between two-way interaction and effectiveness. In [31], only a small r = 0.02 correlation between communication satisfaction and productivity was reported. It is possible that the more long-term, detailed, objective measurement of both communication and productivity in this study allowed the relationship between the two variables which to most is common sense to be more accurately estimated.

The results from Fig 4 indicate the communication graph features played a more important role than speech features in predicting the dependent variables. Among the top graph features, algebraic connectivity, number of nodes and average neighbor degree signify the total number of employees and frequency of interactions between them and graph energy and graph spectrum tells us about the structural complexity of the network. From the speech features, the mean MFCC coefficients are likely tapping into the number of speakers in the graph; the spectral and the energy variability features are likely measuring the number of speakers and frequency of interactions. It is interesting that the fundamental frequency variability is a measure of productivity. This could be a proxy for gender diversity in the organizational unit, although this most certainly requires additional study.

It is important to note that while this study reveals some relationship between communication and productivity, it does not mean that this relationship is causal. It is unknowable from out data whether it is the productivity that induces a change in the network or whether the network induces a change in productivity.

The method described in this paper makes it possible to convert audio-recordings among members of an organization into communication network measures. As such it should be useful to group researchers, who often record all members of a group, and to those organizational researchers who record an entire unit or organization. While the data requirements for the method are demanding, it yields a much more accurate and potentially more valid measure of communication networks than do currently utilized questionnaire methods.

The best choice of a productivity measure can be argued here. Both changed and inserted lines of codes are important measures that cannot be neglected, when it comes to programmer productivity. The inclusion of deleted lines of codes is debatable, as those can be errors or bugs in previously-written codes, that can said to be counter-productive. But at the same time, it can argued that deletion mean shortening of code or making it more compact using improved logic, which is an important aspect of productivity.

This study is unique in terms of organizational communication as it involves long-term, objective, quantitative analysis showing the relationship between a human communication network and productivity in an organization. We have used speech recordings from employees in a software organization to estimate communication networks and extract speech features over a period of 3 years. Effective lines of code was used as the measure of productivity which we attempted to predict using both communication network and speech features. It was found that there exists a moderate relationship between communication and productivity in an organization and it depends on the number of employees, the frequency of conversation between them and the topography of the network. Further investigation can be done by including other forms of communication like, email, texts etc. Besides that, more complex graphs with multiple modules (employee, project, task) can be investigated, which can be a better representative of an organizational setting model. Although, project deadlines were not a prominent feature of SF work because it used extreme programming (XP) as its software development process, it could be interesting to study the communication productivity relationship for different project types and deadline situations. This study does not capture how the communication quantity or speech patterns are affected by specific job stages of a project and how the job stages drive the overall productivity. Since multiple projects overlapped over the whole timeline with employees working on multiple projects at the same time, analyzing various job stages remains a limitation of this study. It requires a more precise analysis of the communication pattern and productivity at various job stages in a project and compare the relationship across various job stages. Furthermore, we can also investigate on productivity on a personal level by analyzing the relationship between communication and productivity for individual employees in the organization.

Supporting information

S1 Appendix. Speech signal features.

(PDF)

Data Availability

Anonymized data is available from openICPSR: https://doi.org/10.3886/E130041V1. Non-anonymized data are not released because they could be used to identify individual participants. Researchers can request access to the non-anonymized data by contacting Dr. Steven Corman (steve.corman@asu.edu), PI of this project, or the ASU IRB (Phone: 480-965-6788 | Fax: 480-965-7772 | Email: research.integrity@asu.edu).

Funding Statement

Dr. Steven Corman NSF PD 11-8031 National Science Foundation https://www.nsf.gov/publications/pub_summ.jsp?ods_key=gpg15001&org=NSF The sponsors played no role in this manuscript.

References

  • 1. Wellman Barry and Berkowitz Stephen D. Social Structures: A Network Approach. Structural analysis in the social sciences 2, 19–61 (1988). [Google Scholar]
  • 2. Pincus J. David. Communication satisfaction, job satisfaction, and job performance. Human Communication Research 12, 395–419 (1986). doi: 10.1111/j.1468-2958.1986.tb00084.x [DOI] [Google Scholar]
  • 3. Mesmer-Magnus Jessica R and DeChurch Leslie. Information sharing and team performance: a meta-analysis. The Journal of applied psychology 94, 535–46 (2009). doi: 10.1037/a0013773 [DOI] [PubMed] [Google Scholar]
  • 4. Kretschmer Tobias. Information and Communication Technologies and Productivity Growth. OECD Digital Economy Papers 195, 1–27 (2012). [Google Scholar]
  • 5. Hellweg Susan A. and Phillips Steven L. Communication and Productivity in Organizations. Public Productivity Review 6, (1982). doi: 10.2307/3380183 [DOI] [Google Scholar]
  • 6. O’Reilly C. A. and Roberts K. H. Task group structure, communication, and effectiveness in three organizations. Journal of Applied Psychology (1977). [Google Scholar]
  • 7. Killworth Peter D. and Bernard H. Russell. Informant Accuracy in Social Network Data. Human Organization 35, 269–286 (1976). doi: 10.17730/humo.35.3.10215j2m359266n2 [DOI] [Google Scholar]
  • 8. Russell Bernard H. and Killworth Peter and Kronenfeld David and Sailer Lee. The Problem of Informant Accuracy: The Validity of Retrospective Data. Annual Review of Anthropology 13, 495–517 (1984). doi: 10.1146/annurev.an.13.100184.002431 [DOI] [Google Scholar]
  • 9. Anagnostopoulos Christos Nikolaos and Iliou Theodoros and Giannoukos Ioannis. Features and classifiers for emotion recognition from speech: a survey from 2000 to 2011. Artificial Intelligence Review 43, 155–177 (2012). doi: 10.1007/s10462-012-9368-5 [DOI] [Google Scholar]
  • 10. El Ayadi Moataz and Kamel Mohamed S. and Karray Fakhri. Survey on speech emotion recognition: Features, classification schemes, and databases. Pattern Recognition 44, 572–587 (2011). doi: 10.1016/j.patcog.2010.09.020 [DOI] [Google Scholar]
  • 11. Batliner Anton and Schuller Bjoern and Seppi Dino and Steidl Stefan and Devillers Laurence and Vidrascu Laurence, et al. The Automatic Recognition of Emotions in Speech. Administrative Science Quarterly (2011). [Google Scholar]
  • 12. Benzeghiba M. and De Mori R. and Deroo O. and Dupont S. and Erbes T. and Jouvet D. et al. Automatic speech recognition and speech variability: A review. Speech Communication 49, 763–786 (2007). doi: 10.1016/j.specom.2007.02.006 [DOI] [Google Scholar]
  • 13. Altun Halis and Polat Gökhan. Boosting selection of speech related features to improve performance of multi-class SVMs in emotion detection. Expert Systems with Applications 36, 8197–8203 (2009). doi: 10.1016/j.eswa.2008.10.005 [DOI] [Google Scholar]
  • 14. Ververidis Dimitrios and Kotropoulos Constantine. Emotional speech recognition: Resources, features, and methods. Speech Communication 48, 1162–1181 (2006). doi: 10.1016/j.specom.2006.04.003 [DOI] [Google Scholar]
  • 15. Tamulevicius Gintautas and Liogiene Tatjana. Low-Order Multi-Level Features for Speech Emotion Recognition. Baltic Journal of Modern Computing 3, 234–247 (2015). [Google Scholar]
  • 16. Koolagudi Shashidhar G. and Rao K. Sreenivasa. Emotion recognition from speech: a review. International Journal of Speech Technology 15, 99–117 (2012). doi: 10.1007/s10772-012-9139-3 [DOI] [Google Scholar]
  • 17. Rong Jia and Li Gang and Phoebe Chen Yi-Ping. Acoustic feature selection for automatic emotion recognition from speech. Information Processing & Management 45, 315–328 (2009). doi: 10.1016/j.ipm.2008.09.003 [DOI] [Google Scholar]
  • 18. Simpson Joseph R. and Snyder Abraham Z. and Gusnard Debra A. and Raichle Marcus E. Emotion-induced changes in human medial prefrontal cortex: I. During cognitive task performance. Proceedings of the National Academy of Sciences 98, 683–687 (2001). doi: 10.1073/pnas.98.2.683 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Bechara Antoine and Damasio Hanna and Damasio Antonio R. Emotion, Decision Making and the Orbitofrontal Cortex. Cerebral Cortex 10, 295–307 (2000). doi: 10.1093/cercor/10.3.295 [DOI] [PubMed] [Google Scholar]
  • 20. Kluemper Donald H. and DeGroot Timothy and Choi Sungwon. Emotion Management Ability: Predicting Task Performance, Citizenship, and Deviance. Journal of Management 39, 878–905 (2013). doi: 10.1177/0149206311407326 [DOI] [Google Scholar]
  • 21. Ashforth Blake E. and Humphrey Ronald H. Emotion in the Workplace: A Reappraisal. Human Relations 48, 97–125 (1995). doi: 10.1177/001872679504800201 [DOI] [Google Scholar]
  • 22. Dobre Ovidiu Iliuta. Employee motivation and organizational performance. Review of Applied Socio-Economic Research 5, 53–60 (2013). [Google Scholar]
  • 23. Ogbonna Emmanuel and Harris Lloyd. Leadership Style, Organizational Culture and Performance: Empirical Evidence from U.K. Companies. International Journal of Human Resource Management—INT J HUM RESOUR MANAG 11, (2000). [Google Scholar]
  • 24. Austin J. R. Transactive memory in organizational groups: The effects of content, consensus, specialization, and accuracy on group performance. Journal of Applied Psychology 88, 866–878 (2003). doi: 10.1037/0021-9010.88.5.866 [DOI] [PubMed] [Google Scholar]
  • 25. Barsade Sigal G. The Ripple Effect: Emotional Contagion and its Influence on Group Behavior. Administrative Science Quarterly 47, 644–675 (2002). doi: 10.2307/3094912 [DOI] [Google Scholar]
  • 26. Corman Steven R., Steiner Elena, Proulx Jeffrey D., Dutta Arindam, Yahja Alex, Scott Poole M., et al. Revisiting the accuracy problem in network analysis using a unique dataset. Social Networks 66, 1–9 (2021). doi: 10.1016/j.socnet.2020.12.010 [DOI] [Google Scholar]
  • 27. Mintzberg Henry. Mintzberg On Management: Inside Our Strange World of Organizations. New York: London: Free Press; (1989). [Google Scholar]
  • 28. Sudhakar Purna and Farooq Ayesha and Patnaik Sanghamitra. Measuring productivity of software development teams. Serbian Journal of Management 7, 65–75 (2012). doi: 10.5937/sjm1201065S [DOI] [Google Scholar]
  • 29. Wagner Stefan and Ruhe Melanie. A Systematic Review of Productivity Factors in Software Development. Software Productivity Analysis and Cost Estimation (SPACE 2008), 1–6 (2008). [Google Scholar]
  • 30. Hall Mark and Frank Eibe and Holmes Geoffrey and Pfahringer Bernhard and Reutemann Peter and Witten Ian H. The WEKA data mining software: an update. SIGKDD Explorations 11, 10–18 (2009). doi: 10.1145/1656274.1656278 [DOI] [Google Scholar]
  • 31.G. T. Goodnight and D. R. Crary and V. W. Balthrop and M. Hazen. The Problem of Informant Accuracy: The Validity of Retrospective Data. International Communication Association annual meeting (1974).

Decision Letter 0

Nersisson Ruban

15 Feb 2021

PONE-D-20-41110

Analyzing the relationship between productivity and human communication in an organizational setting

PLOS ONE

Dear Dr. Arindam Dutta,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Read the reviewer comments carefully and provide detailed answer for every query individually and also incorporate required changes in the manuscript.

Please submit your revised manuscript by 28th Feb 2021. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Nersisson Ruban, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

3. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match.

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: I Don't Know

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This paper uses very detailed voice recording data from what I think is a kind of student software consulting organization at a University over the course of almost 3 years to study whether measures of voice communication can predict productivity. The voice data used in the study is very unique in how complete it is and how long a time period it covers and it allows the authors to generate measures of communication between pairs of individuals inside the organization (employees and clients). The authors find that speech and communication network characteristics can predict software programmer productivity.

I do not have a sufficient understanding of voice measures to comment on the validity what seems to be heroic work by the authors to generate useful measures of communication networks and speech characteristics.

While I think a better understanding of how communication relates to productivity in organizations is a very important question, I am not convinced this study advances our understanding of this question. There are several reasons for this, which I separate out below.

1. The authors do not provide sufficient information about the organization they are studying, except that the work being performed by employees is software programming. The authors argue that the type of work being performed should affect the relationship between communication and productivity, which I agree with, but so should, for instance, the organizational structure, and the sales or work cycles. From what I can gather on my own, it seems as though the software factory is a student run software consultancy which suggests a very flat organizational structure in which most conversations may be between “equals”. It also suggests that work occurs when there is a project available, and otherwise no work occurs. Without understanding how the organization functions and how often work occurs in the organization, it is difficult to know the extent to which the findings generalize to other organizations.

Moreover, the authors mention that their voice data includes conversations between programmers and between programmers and clients. It seems to me like the type of communication that is desirable differs depending on whether a programmer is speaking with a fellow programmer or with a client. I may have missed a discussion of whether the authors drop the client-facing conversations, but I was surprised this wasn’t emphasized more. Given that the measure of productivity used in the paper does not directly capture how satisfied a client was or how big a contract the programmers got, I don’t think the conversations with clients should be included in the study.

2. As the authors point out, the study is not able to determine whether different types of communication patterns cause more or less productivity, but rather whether they can be useful to predict how well the organization is doing (in this case, how productive workers will be). For instance, communication frequency may be a proxy for how excited or engaged workers are in the job or how frustrated they are with the client’s demands. If speech and communication network characteristics are easier to observe and measure than other things like employee attitudes, that using them to predict productivity seems like it could be a very useful tool for employers to improve performance by helping them catch potential problems before they become disasters.

However, I do not think the authors did enough to explain how feasible or practical it is for employers to adopt this in practice. I imagine many employees would object to being recorded while working, and it may be hard to retain high quality employees if they were required to record themselves while at work. Given the importance of communication frequency for predicting productivity, partial recordings may not be sufficient. Generating useful data out of the voice recordings would also be expensive time-wise. Thus, it is unclear to me what the practical takeaways of the paper are.

3. I don’t think my previous comment would be important if the study improved our understanding of why people communicate the way they do when they are more or less productive. As I mention above, there are many reasons why communication patterns may change at the same time as productivity changes but the authors do not tell us anything about these potential mechanisms. I think the authors could do more with their data in this regard. For instance, perhaps they could get data on when project deadlines are and test whether the relationship between communication out productivity changes when workers are facing more urgency. Similarly, perhaps the authors could get information on when students are working on different types of jobs (e.g. maybe students participated in hack-a-thons at some points that were not contracted for). These types of tests may move us forward in our understanding of how speech and communication patterns differ depending on the circumstances, and how those differences relate to productivity.

4. This is a more minor comment, but I wondered if employees would choose to delete or not turn in recordings of less appropriate discussions, including disagreements. If this is possible, perhaps this might explain the lack of findings on speech signal features? Some discussion of possible measurement error of these features might be helpful for understanding how much we can take away from the findings.

5. I am also a bit concerned about mismeasurement of productivity in the lines of code changed may not capture code quality. The authors acknowledge that deletions may or may not improve code quality, but on the converse, additions may or may not improve code quality. At the extreme, if one programmer is particularly frustrated with a co-worker or client, they may write a number of confusing lines that could make it hard for subsequent programmers to successfully edit the code. There are some relatively standard measures of code quality that the authors could have external programmers evaluate the organization’s code on (e.g. complexity, coupling).

Reviewer #2: I think this is great. I like the analyses to measure communication among this group of employees. The sample size is a bit small for my taste, but overall it's quite sound. I think this could be an important paper that change the way we measure communication in the future.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Jul 14;16(7):e0250301. doi: 10.1371/journal.pone.0250301.r002

Author response to Decision Letter 0


4 Mar 2021

We would like to sincerely thank the reviewers for their valuable suggestions. All the suggestions were very helpful, and we have tried our best to work on those suggestions and made necessary changes. The response to all the comments made by the reviewers have been addressed in the 'Response to Reviewers' letter, which also reflects the changes made in the paper.

Attachment

Submitted filename: Response to Reviewers.pdf

Decision Letter 1

Nersisson Ruban

17 Mar 2021

PONE-D-20-41110R1

Analyzing the relationship between productivity and human communication in an organizational setting

PLOS ONE

Dear Dr. Dutta,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Dear author, please address the concern raised by reviewer-1.

Please submit your revised manuscript by March 30 2021 11:59PM.. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Nersisson Ruban, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: I Don't Know

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: No

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you very much for your response to my comments. I appreciate the additional details on the work setting, and the inclusion of statements about the extent to which employees could have deleted their conversations and about the methods not being intended for adoption by organizations.

However, I don't think you sufficiently addressed my concern about speech patterns and productivity being both affected by some additional variable (I suggested deadlines, but it could also be the different stages of the task as outlined by the authors in the revised draft). Thus, it may look like certain speech patterns contribute to productivity in the data but in reality, it could be that certain job stages drive certain speech patterns and those same job stages generate different amounts of coding. I think this is an important limitation of the current paper, and at a minimum, should be mentioned in the conclusion as such.

Reviewer #2: As I said before, I think this is an interesting study that could have important contributions to the field. I think it should be published.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Jul 14;16(7):e0250301. doi: 10.1371/journal.pone.0250301.r004

Author response to Decision Letter 1


26 Mar 2021

We would like to sincerely thank the reviewers for their valuable suggestions. All the suggestions were very helpful, and we have tried our best to work on those suggestions and made necessary changes. The response to all the comments made by the reviewers have been addressed below, which also reflects the changes made in the paper.

Attachment

Submitted filename: Response to Reviewers Round2.pdf

Decision Letter 2

Nersisson Ruban

5 Apr 2021

Analyzing the relationship between productivity and human communication in an organizational setting

PONE-D-20-41110R2

Dear Dr. Dutta,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Nersisson Ruban, Ph.D

Academic Editor

PLOS ONE

Acceptance letter

Nersisson Ruban

21 Jun 2021

PONE-D-20-41110R2

Analyzing the relationship between productivity and human communication in an organizational setting

Dear Dr. Dutta:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Nersisson Ruban

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. Speech signal features.

    (PDF)

    Attachment

    Submitted filename: Response to Reviewers.pdf

    Attachment

    Submitted filename: Response to Reviewers Round2.pdf

    Data Availability Statement

    Anonymized data is available from openICPSR: https://doi.org/10.3886/E130041V1. Non-anonymized data are not released because they could be used to identify individual participants. Researchers can request access to the non-anonymized data by contacting Dr. Steven Corman (steve.corman@asu.edu), PI of this project, or the ASU IRB (Phone: 480-965-6788 | Fax: 480-965-7772 | Email: research.integrity@asu.edu).


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES