Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2024 Mar 21;19(3):e0300338. doi: 10.1371/journal.pone.0300338

Simulated operant reflex conditioning environment reveals effects of feedback parameters

Kyoungsoon Kim 1, Ethan Oblak 2, Kathleen Manella 3, James Sulzer 4,*
Editor: Manabu Sakakibara5
PMCID: PMC10956789  PMID: 38512998

Abstract

Operant conditioning of neural activation has been researched for decades in humans and animals. Many theories suggest two parallel learning processes, implicit and explicit. The degree to which feedback affects these processes individually remains to be fully understood and may contribute to a large percentage of non-learners. Our goal is to determine the explicit decision-making processes in response to feedback representing an operant conditioning environment. We developed a simulated operant conditioning environment based on a feedback model of spinal reflex excitability, one of the simplest forms of neural operant conditioning. We isolated the perception of the feedback signal from self-regulation of an explicit unskilled visuomotor task, enabling us to quantitatively examine feedback strategy. Our hypothesis was that feedback type, biological variability, and reward threshold affect operant conditioning performance and operant strategy. Healthy individuals (N = 41) were instructed to play a web application game using keyboard inputs to rotate a virtual knob representative of an operant strategy. The goal was to align the knob with a hidden target. Participants were asked to “down-condition” the amplitude of the virtual feedback signal, which was achieved by placing the knob as close as possible to the hidden target. We varied feedback type (knowledge of performance, knowledge of results), biological variability (low, high), and reward threshold (easy, moderate, difficult) in a factorial design. Parameters were extracted from real operant conditioning data. Our main outcomes were the feedback signal amplitude (performance) and the mean change in dial position (operant strategy). We observed that performance was modulated by variability, while operant strategy was modulated by feedback type. These results show complex relations between fundamental feedback parameters and provide the principles for optimizing neural operant conditioning for non-responders.

Introduction

Operant conditioning is a commonly used procedure that provides a reinforcing stimulus as a consequence to a desired behavior [1]. Operant conditioning of neural activity has been investigated for decades [2, 3], primarily at the level of the brain [4]. In this manner, an individual learns to self-modulate a neural circuit, with the implication of neuroplastic changes [47]. LaCroix [8] hypothesized that in operant conditioning there are both implicit automatic learning processes occurring in parallel with explicit ones. However, these processes combined with physiology of a complex brain circuit is a poorly understood process [4]. As opposed to complex neural circuity, there exists a simpler model for neurofeedback that targets well-understood monosynaptic stretch reflexes [911]. In this model, an evoked response from electrical stimulation of the peripheral nerve is measured from the innervated muscle (i.e. H-reflex [12]), and its amplitude is directly provided as feedback to the user. This technique, known as operant H-reflex conditioning, was developed by Wolpaw and colleagues both in human and animal models [13, 14]. These researchers have shown the importance of corticospinal tracts in enabling operant learning of the H-reflex amplitude [15], indicated the sites of neuroplastic changes on a synaptic level [16], and provided evidence for its translation to humans [14]. They suggest that operant H-reflex conditioning provides an excellent model for learning of a simple motor skill [13, 17].

Operant H-reflex conditioning typically requires about three months of training. However, there are a substantial portion of individuals who do not learn to self-modulate the spinal circuit, known as non-responders [4, 6]. Such an investment in training for an unknown outcome becomes prohibitive for both scientific research and clinical translation, thus it is critical to understand why certain individuals have difficulty in learning. Operant H-reflex conditioning has traditionally been assumed to rely on implicit mechanisms because of its initiation in rodent models [9, 18, 19]. Further, explicit self-modulation strategies reported in earlier operant H-reflex conditioning work were unrelated to CNS processes responsible for the task-dependent changes in H-reflex size [14]. In contrast, functional magnetic resonance imaging (fMRI) neurofeedback was initiated in humans and was first believed to be entirely governed by explicit processes [7]. However, a contemporary work showed implicit learning with no instructed explicit strategy was successful in modulating early visual cortical activity [20], and further evidence solidified that implicit learning mechanisms were indeed feasible [21]. The relative roles of implicit and explicit processes in operant conditioning of neural signals is still debated [4, 20, 22], and may be influenced by the trained neural substrate. Our anecdotal evidence suggests that even though prior to training we have no conscious ability to regulate monosynaptic reflex activity, explicit mechanisms may play a role in operant H-reflex conditioning [23]. Specifically, we observed the performance of a post-stroke individual trained to operantly condition the H-reflex of the rectus femoris (RF) remained static until the participant was consistently reminded of the instructions.

We have developed a paradigm that helps isolate explicit and implicit learning mechanisms in neural operant conditioning. We previously developed a simulated fMRI neurofeedback environment that separated the ability to self-regulate the neurofeedback signal from its perception by using an explicit, unskilled visuomotor task [24]. We then validated this model empirically in a fMRI neurofeedback experiment [25]. By creating artificial brain activity based on real data and testing outside the scanner, we could vary certain parameters such as hemodynamic delay and feedback delay to efficiently examine the effects of feedback on learning. We found that the feedback timing and hemodynamics affected strategy and performance. These results critically emphasize the importance of understanding the characteristics of the feedback signal that modulate explicit strategy and ultimately operant conditioning performance.

The goal of this study was to further understand the explicit aspect of learning during the operant H-reflex conditioning process. As in our previous simulated neurofeedback paradigm [24], operant conditioning can be broken down into perception, decision making, conditioning ability, and biological variability (noise) components. Based on our data of operant H-reflex conditioning [23], we can simulate the level to which individuals can condition the H-reflex, as well as the range of variability of the H-reflex. Assuming unimpeded perception, we can isolate the effects of simulated feedback on explicit decision-making processes likely occurring during actual neural operant conditioning. Our hypothesis was that several factors likely affected explicit learning: 1) Kr feedback will improve performance compared to Kp feedback [26], 2) larger biological variability will result in worse performance, and 3) the reward threshold modulates performance and strategy.

During this cognitive experiment, neurologically intact participants were asked to play a simple computer game, wherein the task was to rotate a virtual rotary knob representative of an operant strategy. The simulated H-reflex was composed of a knob-controllable canonical form of the H-reflex with biological variability and measurement signal noise extracted from the actual H-reflex data collected previously [23]. As the inherently variable nature of H-reflex amplitude has been shown to be correlated with the fluctuations in motoneuronal membrane potential [27], application of data-driven biological variability and signal noise enabled our simulation environment to be as similar to the neurophysiological environment of operant H-reflex conditioning as possible. The goal was to align the knob with a hidden target, of which the proximity of the knob to the target was proportional to the feedback signal. In other words, participants were asked to “down-condition” the virtual feedback amplitude, which was achieved by placing the knob as close as possible to the hidden target. The task is analogous to a tuning dial on a radio, where the participant aims to find the best signal. Thus, we are replacing a skilled operant conditioning task with a simplified, unskilled tuning task that enables us to quantify operant strategies at an unprecedented level. We then compared this model to real operant H-reflex conditioning performance to demonstrate the similarity in learning processes between the simulation environment and that of operant H-reflex conditioning. This experiment represents a novel approach to understanding the role of feedback parameters in explicit learning that is likely taking place during neural operant conditioning. The simulated operant H-reflex conditioning environment provides an efficient method for analyzing and possibly improving learning. This approach may also assist in identifying non-responders, ultimately enhancing the robustness of operant H-reflex conditioning.

Methods

Subjects

A total of 41 healthy participants were recruited with no history of vision impairment, cognitive impairment, and neurological disease or injury. The participants were 24 men and 17 women aged 20–41 between April and July 2021. The study was considered exempt by the University of Texas at Austin Institutional Review Board. While consent was not required, it was acquired via email or verbally from all participants prior to participation. There was no documentation collected linking participant identity to their data. To prevent variation in one’s concentration level, sessions were performed at the same time of the day for everyone. The experiment consisted of three sessions, Sessions 1 and 2 took approximately 1.5 hours to complete and Session 3 took approximately 30 minutes. All sessions were performed at least a day apart.

Experimental protocol and data collection

This experiment was conducted virtually, where the participant was asked to use their personal computer and play a simple computer game via a web application (web app). The web app was designed using the Matlab App Designer (MathWorks, Natick, MA), and was established at the University of Texas at Austin’s Matlab Web App server. Participants were emailed a link prior to the experiment. The participant and investigator met virtually on the scheduled date, where a webcam was used to monitor the session to ensure attention and efficient troubleshooting in case of any problems. The experimental protocol consisted of 4 stages: instruction, consent, demonstration (practice run), and the experiment. The experiment was carried out over three separate sessions at least a day apart.

During the Instruction stage, participants were briefly informed about the purpose of the study, approximate duration, future experimental schedule, and compensation for their participation. The virtual consent was held as a reconfirmation of the participants’ voluntary commitment toward the task and individuals were free to withdraw from the experiment whenever desired. If the participant chose to participate in the study, the participant provided their full name and e-mail address as an electrical signature. Personal information acquired at this stage was encrypted and data was anonymized based on this encryption.

The Demonstration stage was considered a practice run and participants were informed of the visual layout, the keyboard control, task goal, and the experimental structure of the experiment in a detailed manner. The visual layout consisted of rotary dial knob, feedback bar, attempt number, and cumulative success rate (Fig 1). The layout was modeled after the traditional protocol in earlier work [14]. The rotation of the rotary knob was controlled by keyboard inputs, arranged based on desired hand: fast counterclockwise (A if left side desired or L if right side desired), slow counterclockwise (S, K), fast clockwise (F, H), slow clockwise (D, J), and select (spacebar). The task goal was to align the rotary knob with a hidden target (invisible during experiment but shown in orange in Fig 1). The displacement of the rotary knob is a one-dimensional reduction serving as a proxy for the decision making occurring during operant H-reflex conditioning. That is, the amount of angular displacement of the rotary knob is reflective of the mental strategy selection or effort during real operant H-reflex conditioning. The dial was programed to move after each keystroke and did not move continuously when the key was pressed and held. There was no time limit for each trial, and participants were allowed to move the dial as much as necessary, before confirming their decision by pressing the spacebar. Once the position was selected the web app generated a simulated H-reflex.

Fig 1. Visual layout of the web application and feedback parameter.

Fig 1

(a) A feedback screen for the participant, in which they were asked to rotate the virtual rotary knob to find the hidden target. The feedback bar indicated the amount of error (Knowledge of Performance, Kp feedback) or success/failure information (Knowledge of Results, Kr feedback) in finding the target via changing the bar height (Kp) or color (Kr). Participants received additional feedback of a running score of their performance and trial number. (b) Visualization of feedback parameters: feedback type (performance, Kp, knowledge Kr, and both KpKr), biological variability (low, LV and high, HV), and reward threshold (easy, ET, moderate, MT, and difficult, DT).

Simulated H-reflex

We modeled the H-reflex (Fig 2) by decomposing it into parts to investigate the role of these subcomponents on learning in a simulation. For each trial (k), when the participant explored the hidden target and confirmed decision by pressing spacebar, a simulated H-reflex (hsim(k)) was generated and presented as feedback bar graph (Fig 3).

Fig 2. Representation of rectus femoris (RF) H-reflex.

Fig 2

Electromyography (EMG) activity during femoral nerve stimulation is depicted. After the onset of electrical stimulation on the femoral nerve, a motor response (M-wave) and monosynaptic spinal reflex (H-reflex) is elicited. The H-reflex profile was decomposed to investigate the role of these subcomponents on learning in a simulation.

Fig 3. Simulated H-reflex.

Fig 3

Simulated H-reflex (hsim) is generated by applying decision gain (ac), biological variability (β), and noise (aN) on the H-reflex time-course and the peak-to-peak magnitude is provided as the visual feedback for the participant. Participant uses information to adjust one’s strategy (Δθ) to either minimize the feedback bar height or turn the bar green.

The simulated H-reflex was composed of knob-controllable canonical form of the H-reflex, decision gain, biological variability and signal noise driven from the actual H-reflex data. The canonical form was a single period sine function. The amplitude of the sine function ac(k) was determined by the difference (error) between the dial position (θ(k)) and the hidden target (θtar), and multiplied by a predetermined gain, g:

ac(k)=g(θ(k)θtar) (1)

The peak-to-peak value of the canonical form was a linear function set at 1 when the error was 180° and 0.5 when the error was 0°. Thus, for down-conditioning, the best possible performance of 0.5 was extracted empirically from earlier work on operant down-conditioning of the RF H-reflex [23]. In other words, having a peak-to-peak value closer to 0.5 implied the subject was successful in matching the hidden target, which can be translated as success in the context of down-conditioning, whereas value closer to 1 implied failure. We modeled the natural biological variability of the H-reflex (σ2) with a normal distribution obtained from a set of participants’ H-reflexes, also from previous work [23]. All analyses and distribution model fitting was performed using MATLAB software (MATLAB 2019a, Mathworks, Natick, MA). We used the minimum and maximum variability (LV, σ2 = 0.25 and HV, σ2 = 0.75, respectively) of the dataset. This variability was applied in the form of normal distribution (β(k)~N(μ,σ2)) generated by Box-Muller transform [28]. Based on the same distribution, the dominant noise power (aN) of H-reflex signals was analyzed using Fast Fourier Transform (FFT) and applied to the simulated H-reflex. To measure performance, we acquired the simulated H-reflex, hsim(k), using the following equation:

hsim(k)=(sin(t)*ac(k)*β(k)+aN)pkpk (2)

For the Demonstration, the participant conducted 5 familiarization trials for two conditions, Kp, and Kr. Both terms are commonly used in psychology [26], where Kp provides information about how the task was achieved and Kr is knowledge about whether the goal of the task was achieved. In this sense, Kp focused on reducing the feedback bar height and Kr focused on maintaining the bar height below a threshold. During Kp, a blue bar was presented (Fig 1B, top right), in which the height was peak-to-peak magnitude of the simulated H-reflex. No bar color change was associated with Kp feedback. For Kr, the bar color changed based on the result of the comparison between the simulated H-reflex magnitude and the pre-determined threshold levels (Fig 1B, middle right). The threshold levels were chosen as the 44th, 66th, and 77th percentile of the 1,350 evoked H-reflexes without feedback [23], designated as easy/moderate/difficult thresholds (ET = 0.46, MT = 0.7 and DT = 0.96, respectively). If the magnitude of the simulated H-reflex was below the threshold, the bar turned green indicating success and the cumulative score increased. If hsim(k) was larger than the threshold, the bar turned red, indicating failure and the score decreased. The bar height and the threshold were kept hidden from the participants during the Kr condition and only the binary result with fixed bar height was provided as feedback. Not introduced in the Demonstration, but used later in the Experiment, is the combination of Kp and Kr feedback, which is the KpKr condition, but with a moving bar height as described above (Fig 1B, bottom right), reflecting current operant H-reflex conditioning practices [14].

Following the Demonstration, we introduced the Experiment phase. The experimental structure comprised of 10 different conditions including factors of three feedback types (Kp/Kr/KpKr), low/high biological variability (LV/HV), and three reward threshold levels (ET/MT/DT), the latter factor during Kr feedback only (Fig 4A).

Fig 4. Experimental condition and protocol.

Fig 4

(a) Total of ten conditions were tested, based on the different feedback types (performance, Kp, results, Kr, and both, KpKr), biological variability (low, LV and high, HV), and reward threshold (easy, ET, moderate, MT, and difficult, DT). (b) There were three sessions in this experiment. Each session was comprised of 10 runs. Within each run, there were 4 conditions for Session 1 and 2, and Session 3 had two conditions. For each condition, the participant was given 35 trials.

We conducted 3 sessions on different days at least one day apart: Session 1 (C1-C4), Session 2 (C5-C8), and Session 3 (C9-C10) (Fig 4B). The Experiment stage consisted of 10 runs of each condition of 35 trials each (Fig 4B). A “trial” was defined as a single decision, where the individual explored the hidden target using keyboard inputs, confirmed one’s decision by pressing spacebar, and feedback bar was updated accordingly. The order of the conditions within a run was pseudo-randomized to avoid effects of ordering [29]. Each session took approximately 1–1.5hrs and the participant was able to take an optional short break in between runs with a mandatory 1-minute break after the 5th run. The subject was informed of the progress in percentage by a pop-up window after each even numbered run (i.e., 2nd, 4th, 6th, and 8th run).

Statistical analysis

Our main outcome measures consisted of the mean simulated H-reflex amplitude (performance, hsim(k)) and the mean change in dial position between trials (operant strategy, Δθ(k)). Operant strategy is a measure of the participant’s exploration in finding a hidden target, which was the angular difference between the present dial position and previous trial’s dial position. To determine the effects of each of the experimental parameters (feedback type, biological variability, and reward threshold) on the performance and operant strategy of the participants, the average performance and operant strategy of different conditions were compared by one-way repeated measures ANOVA with Tukey HSD post-hoc test (α < 0.05). Our main hypotheses were: 1) Kr feedback will improve performance compared to Kp feedback, 2) larger biological variability will result in worse performance, and 3) the difficult reward threshold (DT) will worsen performance and make the operant strategy less aggressive (or exploratory). In addition, to test for interaction effects of experimental parameters on performance and operant strategy (e.g., interaction between the biological variability and feedback type on performance), two-way repeated measures ANOVA with Tukey HSD post-hoc test was used.

Similarity of learning process during simulated environment: Comparison with real operant H-reflex conditioning

To illustrate similarities between learning in the simulated environment and actual operant H-reflex conditioning behavior, we used a statistical model. We computed a single linear mixed model (LMM) based on conditions C9 and C10 from all participants in the simulated environment (KpKrLVMT and KpKrHVMT). We included the performance mean across a single run as the dependent variable with the fixed effect of biological variability and the random effect of participant. Using this LMM driven from the simulated environment, we then input measured (i.e., real) H-reflex variability extracted from data collected previously for five healthy and two participants post-stroke performing operant RF H-reflex conditioning [23], to calculate the estimated performance. Later, we normalized the estimated performance magnitude for direct comparison with real operant H-reflex conditioning performance, using the method mentioned in our previous study [23]. We compared the estimated performance from the LMM based on measured H-reflex variability with the real operant H-reflex conditioning performance.

There were 24 training sessions, wherein each session was comprised of 3 runs. Using the model, we predicted the performance mean of each of 72 runs. To examine the prediction accuracy, a Pearson’s correlation coefficient with (df = 70, α < 0.05) was computed to assess the linear relationship between the actual performance and the estimated performance.

Results

Effect of biological variability and feedback type on performance

Increased biological variability worsened performance, as evidenced by higher values (Table 1). Under both levels of biological variability, the Kp feedback type exhibited the poorest performance and the KpKr exhibited the best performance (Kp-KpKr = 0.037±0.008, p < 0.0001, Tukey HSD) (F). The difference was larger during the high variability conditions than low variability conditions. We observed a strong trend towards statistical significance of an interaction effect between the biological variability and feedback type on performance (F(2,216) = 2.98, p = 0.05, two-way ANOVA). A summary of comparisons is provided in Fig 5 and Table 1. Detailed values of performance and comparison are provided in S1 Table.

Table 1. Effect of biological variability and feedback type on performance and operant strategy (moderate threshold).

Pair-wise Comparison Performance Operant Strategy
HVLV 0.114±0.004 (***) -1.449±0.074° (ns)
KpKpKr 0.037±0.008 (****) 1.168±0.080° (ns)
KpKr 0.015±0.005 (*) 1.512±0.061° (ns)
KrKpKr 0.021±0.009 (***) -0.344±0.086° (ns)

Values represent overall mean ± SE. Tukey HSD post hoc tests conducted for pair-wise comparisons (↔) between different conditions.

Statistical significance

(* p<0.05

**p<0.01

***p<0.001

****p<0.0001)

Fig 5. Effect of biological variability and feedback type on performance and operant strategy (moderate threshold).

Fig 5

On each box, the notch indicates the median, and the bottom and top edges of the box indicate the 25th and 75th percentiles, respectively. The whiskers extend to the most extreme data points not considered outliers. Red cross (‘+’) marker symbol represents outliers. Grey lines connect each participants’ value across different conditions. Each color represents different feedback types (Blue: Kp, Green: Kr, and Turquoise: KpKr). Results for performance and operant strategy under low and high biological variability conditions are shown. Tukey HSD pair-wise comparisons between different feedback type during each biological variability are added. Statistical significance (* p<0.05, **p<0.01, ***p<0.001, ****p<0.0001).

Effect of biological variability and feedback type on operant strategy

Operant strategy, quantified by the change in angle of the dial (Δθ), was not significantly different across biological variability levels. Also, no significant difference in operant strategy was observed across different feedback types overall (Table 1). However, during LV, operant strategy was more aggressive during Kr compared to Kp feedback (Kp-Kr = 4.79±0.09, p < 0.0001, Tukey HSD) (Fig 5). This inconsistent result raised the question of whether there was any interaction effect between the biological variability and feedback type on operant strategy. Reduced biological variability enhanced the difference in operant strategy between Kp and Kr feedback (F(2,216) = 4.683, p < 0.05, two-way ANOVA). Results are summarized in Fig 5 and Table 1. Detailed values of strategy and comparison are provided in S1 Table.

Effect of biological variability and threshold level on performance

Increased biological variability worsened performance overall (Table 2). At low variability, performance was worst at the easy threshold level compared to difficult and moderate threshold levels (Table 2). However, at high variability, we did not observe any difference in performance across different threshold levels (p>0.05). Using a two-way repeated measures ANOVA, we observed a significant interaction effect between the biological variability and threshold level (F(2,238) = 14.2, p<0.001, two-way ANOVA) on performance. A summary of comparisons is found in Fig 6 and Table 2. Detailed values of performance and comparisons are provided in S2 Table.

Table 2. Effect of biological variability and threshold level on performance and operant strategy.

Pair-wise comparison Performance Operant strategy
HVLV 0.105±0.004 (***) -0.562±0.037 (ns)
DTET -0.040±0.001 (****) 8.772±0.071 (****)
DTMT 0.001±0.001 (ns) 5.281±0.086 (****)
MTET -0.041±0.001 (****) 3.485±0.092 (**)

Values represent overall mean ± SE. Tukey HSD was used for pair-wise comparisons between conditions of high and low variability (HV, LV) and difficult, moderate, and easy threshold (DT, MT, ET).

Statistical significance

(* p<0.05

**p<0.01

***p<0.001

****p<0.0001)

Fig 6. Effect of biological variability and threshold level on performance and operant strategy.

Fig 6

On each box, the notch indicates the median, and the bottom and top edges of the box indicate the 25th and 75th percentiles, respectively. The whiskers extend to the most extreme data points not considered outliers. Red cross (‘+’) marker symbol represents outliers. Grey lines connect each participants’ value across different conditions. Each color represents different reward thresholds (Black: Difficult, Dark gray: Moderate, and Light gray: Easy). Results under low and high biological variability conditions are exhibited. Tukey HSD was used for pair-wise comparisons between conditions. Statistical significance (* p<0.05, **p<0.01, ***p<0.001, ****p<0.0001).

Effect of biological variability and threshold level on operant strategy

Threshold level had a significant effect on operant strategy (F(2,238) = 28.61, p<0.001, two-way ANOVA), with an easier threshold resulting in a less aggressive strategy (Fig 6). We did not observe a significant effect of biological variability on operant strategy (F(1,238) = 0.32, p = 0.5703, one-way ANOVA), but increased signal variance decreased the effect of threshold on strategy (F(2,238) = 6.48, p<0.005, two-way ANOVA) (Fig 6). Pairwise comparisons also indicated increased aggressiveness in operant strategy as threshold became more difficult (Table 2). Detailed values of strategy and comparisons are provided in S2 Table.

Similarity in learning during simulation: Comparison with operant H-reflex conditioning result (run-by-run analysis)

Our simulated results demonstrated fair to strong positive correlations with real operant H-reflex conditioning results [30]. Real performance of 7 participants from previous work (5 healthy and 2 post-stroke) [23] was compared to simulated performance estimated by the LMM (Table 3). All participants’ data were significantly correlated to the simulated data. The strength of the correlations was strong in one healthy individual, moderate in three healthy individuals, and fair in one healthy individual and two post-stroke individuals. A representative example of a strong correlation between real and simulated data is shown in Fig 7. Real data (solid line) consisted of the mean performance of each run of operant H-reflex conditioning training (72 runs = 24 sessions x 3 run/session). Estimated performance based on the LMM using the measured H-reflex variability extracted from each specific run is shown as a dashed line (Fig 7).

Table 3. Comparison of the simulated environment to real operant H-reflex conditioning.

Pearson’s Correlation
Subject R p-value
H1 0.832 < 0.0001
H2 0.676 < 0.0001
H3 0.500 < 0.0001
H4 0.598 < 0.0001
H5 0.491 0.0059
S1 0.308 0.0195
S2 0.411 0.0003

Comparison between two environments comprised of testing a correlation between the performance during real operant H-reflex conditioning for 5 healthy participants (H1-5) and 2 participants with stroke (S1-2) and that of simulated environment.

Statistical significance

(* p<0.05

**p<0.01

***p<0.001

****p<0.0001)

Fig 7. Representative comparison of real and simulated performance.

Fig 7

Normalized performance (%) of H-reflex for simulated environment, real operant H-reflex conditioning environment, and random environment are presented. A simulated environment driven LMM was used to predict real operant H-reflex performance in a single individual (H1) given the biological variability for each run. The real consisted of the mean performance in each of 72 training runs.

Discussion

The goal of the present study was to investigate the effect of feedback type, biological variability, and reward threshold on individuals’ feedback performances and operant strategies in a simulated operant H-reflex conditioning environment. We used a novel simulation environment based on real operant H-reflex conditioning parameters to isolate the effect of feedback parameters on explicit learning. Our main findings, consistent with our hypotheses, were that 1) Kr feedback resulted in better performance than Kp feedback alone, 2) larger biological variability worsened feedback performance, 3) biological variability modulated the effect of feedback type on operant strategy, 4) biological variability modulated the effect of reward threshold on strategy, and 5) a difficult reward threshold resulted in better performance and more aggressive operant strategies. Performance in the simulated environment, albeit governed entirely by explicit learning processes, was found to be similar to actual operant H-reflex conditioning performance. This study is a new approach to understanding the learning mechanisms of operant H-reflex conditioning. Our results suggest that explicit processes play a role in operant H-reflex conditioning, and this process is modulated by feedback parameters. This paradigm can be used to quickly examine different feedback parameters on operant learning and potentially identify non-responders, ultimately improving the procedural robustness of self-modulation of H-reflex activity.

Neural operant conditioning incorporates both explicit and implicit learning processes that are difficult to delineate. In our earlier work, we used a novel simulated operant conditioning environment approach to separate implicit and explicit processes based on real fMRI data [24] and then validated this model experimentally [25]. One of the novelties of this work is the adaptation of this paradigm for operant H-reflex conditioning, a significant development because it allows investigation of basic skill learning principles. Motor skill learning paradigms have different levels of variability, such as the efferent command and variability of the environmental dynamics [31, 32]. Operant H-reflex conditioning has been classified as a simple motor learning task [33]. In this study, we use an unskilled motor task as a proxy for operant strategy and then simulated the natural variability of the H-reflex [34], thereby transferring efferent variability to environmental variability. This innovation allows investigation into two factors: operant strategy unaffected by efferent noise due to the use of an unskilled task as a proxy, and the controlled analysis of the effect of biological variability.

Thus, this approach enables an understanding of the operant strategies one uses. This perspective into operant strategies is unique compared to tasks such as neurofeedback, where operant strategies can only be reported anecdotally, or motor skill learning, where the explicit strategies can be measured, but are filtered by kinematics and muscle activity.

Biological variability played a multifaceted role on performance and strategy. We expected the performance to be worse with higher biological variability. Signal noise affects motor learning [35, 36], and more specifically, has been shown to reduce operant H-reflex conditioning in animal models [19]. The results from the simulated environment aligned with this expectation. Additionally, we expected one’s strategy to be less exploratory or aggressive in case of larger variability given lower reliability on feedback, which was confirmed by our data. We observed that during low variability, Kp feedback resulted in more aggressive operant strategies than Kr feedback, but we did not observe a difference under high variability. Thus, feedback type affects operant strategy only when the variability is low. This observation is novel because we have focused on the interaction effect of biological variability and feedback type on operant strategy, while other studies have only focused on either the biological variability [32, 37] or feedback type [38].

We have also explored the effect of reward threshold level on performance and operant strategy. Previous work in visuomotor learning [39] has emphasized the effect of negative feedback (e.g., failure) on increased variability of subject’s motor strategy during reward-based learning. Also, work in motor learning has indicated that reward threshold affects motor learning [32], and as such, we hypothesized that making the reward threshold more difficult than what is typically provided in operant H-reflex conditioning testing [14] would worsen performance. We observed that performance was affected by threshold level, biological variability, and the interaction between the two. Specifically, at low variability, the easy threshold level worsened performance, however, at high variability, we did not observe a change in performance. Such an interaction is expected as the environment (noise) is known to modulate the effect of task difficulty on motor learning [40]. Given the specific parameters of this task, it would be difficult to compare other work in relation to these findings. However, from an explicit learning perspective, an easier threshold is advised against, when signal noise is sufficiently low.

We observed that at low variability Kr feedback was more effective in improving performance than during Kp feedback. Kr feedback has typically been associated with skilled motor learning processes, in short term and long-term learning [41]. In this study, we show the value of Kr in an unskilled task, independent of any long-term learning. Interestingly, when biological variability was high adding Kp to Kr resulted in further improvement in performance however not when variability was low. Since Kp provides feedback of the distance to target, given low variability, and thus high confidence in the accuracy of the feedback, we would expect Kp to outperform Kr. Within the application of operant reflex conditioning, Kr feedback is sufficient for explicit learning of the task. Given the association of Kr feedback with long-term learning [41, 42] it likely also suffices for implicit processes. However, across biological variability, KpKr feedback showed the best performance. Thus, our results support the continued use of KpKr feedback in operant reflex conditioning.

We found fair to strong correlations between performance in our explicit task and real operant H-reflex conditioning. The amount of variance explained for real operant H-reflex conditioning performance by this purely explicit model suggests that explicit learning likely plays a role in operant H-reflex conditioning. Previous studies [33, 43] have suggested that the process of learning or skill acquisition, such as operant conditioning, involves both explicit and implicit mechanisms. Our simulated environment, however, focused only on the explicit mechanisms (i.e., virtual rotary knob control). Given the growing evidence for the primary role of implicit mechanisms in neural operant conditioning [20, 44, 45], this finding that explicit processes explain so much variance in the data was surprising. Although inconclusive, we observed that the biological variability explained by our model was lowest in the two post-stroke individuals. Despite the difference in correlations with our model, the performance of the two post-stroke individuals during real operant H-reflex conditioning was equal or better than the healthy individuals, perhaps suggesting different learning processes occurring post-stroke.

Our simulation paradigm possesses two major limitations that preclude drawing a direct inference of how learning from simulation may be transferred to real operant H-reflex conditioning capability. First, our simulation model was designed as a one-dimensional study, which solely focused on the angle of the rotary knob with a single global maxima and minima. In the real operant H-reflex conditioning environment, however, reported operant strategy is multi-dimensional and varies within and across participants, which could lead to strategic local maxima and minima [14, 46]. The current paradigm should not show run-by-run learning (Fig 7) because there was no strategy or information in this unskilled task to carry over to the following runs. However, future investigations could examine learning on a trial-by-trial basis [47, 48]. Other approaches could adapt this paradigm to incorporate multi-dimensional strategies. Second, our feedback parameters (e.g., biological variability, reward threshold) were chosen based on a limited pool of 7 subjects’ data from our prior work [23]. As H-reflex is known for its large biological variability originating from various reasons [49], our two-level variability analysis can be expanded to further investigate the effect of multi-level biological variability.

Conclusions

We developed a simulated environment of operant H-reflex conditioning to investigate the effects of feedback parameters on explicit learning. The model explained a large portion of the variance of the real H-reflex conditioning despite lacking any implicit learning process. This simulated paradigm may potentially allow the investigation of parameters beyond the capability of real operant H-reflex conditioning and far more efficiently. Our model suggests that the conditions for best operant H-reflex conditioning performance is to provide both reward and error feedback with at least a moderate (66th percentile success rate) threshold, particularly with low biological variability of the H-reflex. Operant strategy is most aggressive at low variability with error feedback (knowledge of performance) and with a more difficult reward threshold. As variance increases, the effects of feedback type and reward threshold are not as strong. The future of this simulation paradigm may provide a better understanding of learning strategies and ability to identify non-responders, ultimately enhancing the robustness of operant H-reflex conditioning.

Supporting information

S1 Table. Effect of biological variability and feedback type on performance and operant strategy (moderate threshold).

Values represent overall mean ± SE. Tukey HSD was used for pair-wise comparisons (↔) between conditions of high and low variability (HV, LV) and different feedback types (Kp, Kr, KpKr). Statistical significance (* p<0.05, **p<0.01, ***p<0.001, ****p<0.0001).

(PDF)

pone.0300338.s001.pdf (75.5KB, pdf)
S2 Table. Effect of biological variability and reward threshold on performance and operant strategy.

Values represent overall mean ± SE. Tukey HSD was used for pair-wise comparisons (↔) between conditions of high and low variability (HV, LV) and difficult, moderate, and easy threshold (DT, MT, ET). Statistical significance (* p<0.05, **p<0.01, ***p<0.001, ****p<0.0001).

(PDF)

pone.0300338.s002.pdf (91.6KB, pdf)

Data Availability

Data is available at: Sulzer, James, and Kyoungsoon Kim. 2023. “Operant Conditioning Simulation Paradigm.” OSF. May 26. DOI 10.17605/OSF.IO/Y8P47 https://osf.io/y8p47/.

Funding Statement

This work was financially supported in part by the NIH/NICHD (P2CHD086844, Kautz), and JS is the recipient. This work was also supported by the NICHD under the National Institutes of Health under the Award Number R01HD100416. The contents are solely the responsibility of the authors and do not necessarily represent the official views of the NIH or NICHD. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Skinner BF. The encyclopedia of educaiton. New York: Macmillan and Free Press, 1971. [Google Scholar]
  • 2.Kamiya J. The First Communications About Operant Conditioning of the EEG. J Neurother 2011; 15: 65–73. [Google Scholar]
  • 3.Fetz E. Operant Conditioning of Cortical Unit Activity. Sci New Ser 1969; 163: 955–958. doi: 10.1126/science.163.3870.955 [DOI] [PubMed] [Google Scholar]
  • 4.Sitaram R, Ros T, Stoeckel L, et al. Closed-loop brain training: the science of neurofeedback. Nat Rev Neurosci 2017; 18: 86–100. doi: 10.1038/nrn.2016.164 [DOI] [PubMed] [Google Scholar]
  • 5.Blefari ML, Sulzer J, Hepp-Reymond M-C, et al. Improvement in precision grip force control with self-modulation of primary motor cortex during motor imagery. Front Behav Neurosci; 9. Epub ahead of print 13 February 2015. doi: 10.3389/fnbeh.2015.00018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Sulzer J, Haller S, Scharnowski F, et al. Real-time fMRI neurofeedback: Progress and challenges. NeuroImage 2013; 76: 386–399. doi: 10.1016/j.neuroimage.2013.03.033 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Sulzer J, Sitaram R, Blefari ML, et al. Neurofeedback-mediated self-regulation of the dopaminergic midbrain. NeuroImage 2013; 83: 817–825. doi: 10.1016/j.neuroimage.2013.05.115 [DOI] [PubMed] [Google Scholar]
  • 8.Jm L. The acquisition of autonomic control through biofeedback: the case against an afferent process and a two-process alternative. Psychophysiology; 18. Epub ahead of print September 1981. doi: 10.1111/j.1469-8986.1981.tb01828.x [DOI] [PubMed] [Google Scholar]
  • 9.Wolpaw JR. Operant conditioning of primate spinal reflexes: the H-reflex. J Neurophysiol 1987; 57: 443–459. doi: 10.1152/jn.1987.57.2.443 [DOI] [PubMed] [Google Scholar]
  • 10.Wolf SL, Segal RL. Conditioning of the Spinal Stretch Reflex: Implications for Rehabilitation. Phys Ther 1990; 70: 652–656. doi: 10.1093/ptj/70.10.652 [DOI] [PubMed] [Google Scholar]
  • 11.Mrachacz-Kersting N, Kersting UG, de Brito Silva P, et al. Acquisition of a simple motor skill: task-dependent adaptation and long-term changes in the human soleus stretch reflex. J Neurophysiol 2019; 122: 435–446. doi: 10.1152/jn.00211.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Palmieri RM, Ingersoll CD, Hoffman MA. The Hoffmann Reflex: Methodologic Considerations and Applications for Use in Sports Medicine and Athletic Training Research. 10. [PMC free article] [PubMed] [Google Scholar]
  • 13.Thompson AK, Wolpaw JR. The Simplest Motor Skill: Mechanisms and Applications of Reflex Operant Conditioning. Exerc Sport Sci Rev 2014; 42: 82–90. doi: 10.1249/JES.0000000000000010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Thompson AK, Chen XY, Wolpaw JR. Acquisition of a Simple Motor Skill: Task-Dependent Adaptation Plus Long-Term Change in the Human Soleus H-Reflex. J Neurosci 2009; 29: 5784–5792. doi: 10.1523/JNEUROSCI.4326-08.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Chen XY, Carp JS, Chen L, et al. Corticospinal tract transection prevents operantly conditioned H-reflex increase in rats. Exp Brain Res 2002; 144: 88–94. doi: 10.1007/s00221-002-1026-8 [DOI] [PubMed] [Google Scholar]
  • 16.Wolpaw J, Chen XY. Operant Conditioning of Reflexes. Encycl Neurosci 2009; 7: 225–233. [Google Scholar]
  • 17.Norton JJ, Wolpaw JR. Acquisition, maintenance, and therapeutic use of a simple motor skill. Curr Opin Behav Sci 2018; 20: 138–144. doi: 10.1016/j.cobeha.2017.12.021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Chen XY, Wolpaw JR. Operant conditioning of H-reflex in freely moving rats. J Neurophysiol 1995; 73: 411–415. doi: 10.1152/jn.1995.73.1.411 [DOI] [PubMed] [Google Scholar]
  • 19.Carp JS, Tennissen AM, Chen XY, et al. H-Reflex Operant Conditioning in Mice. J Neurophysiol 2006; 96: 1718–1727. doi: 10.1152/jn.00470.2006 [DOI] [PubMed] [Google Scholar]
  • 20.Shibata K, Watanabe T, Sasaki Y, et al. Perceptual Learning Incepted by Decoded fMRI Neurofeedback Without Stimulus Presentation. Science 2011; 334: 1413–1415. doi: 10.1126/science.1212003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Sepulveda P, Sitaram R, Rana M, et al. How feedback, motor imagery, and reward influence brain self‐regulation using real‐time fMRI. Hum Brain Mapp 2016; 37: 3153–3171. doi: 10.1002/hbm.23228 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Muñoz-Moldes S, Cleeremans A. Delineating implicit and explicit processes in neurofeedback learning. Neurosci Biobehav Rev 2020; 118: 681–688. doi: 10.1016/j.neubiorev.2020.09.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Kim K, Akbas T, Lee R, et al. Self-modulation of rectus femoris reflex excitability in humans. Sci Rep 2023; 13: 8134. doi: 10.1038/s41598-023-34709-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Oblak EF, Lewis-Peacock JA, Sulzer JS. Self-regulation strategy, feedback timing and hemodynamic properties modulate learning in a simulated fMRI neurofeedback environment. PLOS Comput Biol 2017; 13: e1005681. doi: 10.1371/journal.pcbi.1005681 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Oblak EF, Sulzer JS, Lewis-Peacock JA. A simulation-based approach to improve decoded neurofeedback performance. NeuroImage 2019; 195: 300–310. doi: 10.1016/j.neuroimage.2019.03.062 [DOI] [PubMed] [Google Scholar]
  • 26.Sharma DA, Chevidikunnan MF, Khan FR, et al. Effectiveness of knowledge of result and knowledge of performance in the learning of a skilled motor activity by healthy young adults. J Phys Ther Sci 2016; 28: 1482–1486. doi: 10.1589/jpts.28.1482 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Gossard JP, Floeter MK, Kawai Y, et al. Fluctuations of excitability in the monosynaptic reflex pathway to lumbar motoneurons in the cat. J Neurophysiol 1994; 72: 1227–1239. doi: 10.1152/jn.1994.72.3.1227 [DOI] [PubMed] [Google Scholar]
  • 28.Box GEP, Muller ME. A Note on the Generation of Random Normal Deviates. Ann Math Stat 1958; 29: 610–611. [Google Scholar]
  • 29.van Casteren M, Davis MH. Mix, a program for pseudorandomization. Behav Res Methods 2006; 38: 584–589. doi: 10.3758/bf03193889 [DOI] [PubMed] [Google Scholar]
  • 30.Portney LG. Foundations of Clinical Research: Applications to Evidence-Based Practice. 4th ed. DAVIS F.A., 2020. [Google Scholar]
  • 31.Maurice P, Hogan N, Sternad D. Predictability, force, and (anti)resonance in complex object control. J Neurophysiol 2018; 120: 765–780. doi: 10.1152/jn.00918.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Huber ME, Kuznetsov N, Sternad D. Persistence of reduced neuromotor noise in long-term motor skill learning. J Neurophysiol 2016; 116: 2922–2935. doi: 10.1152/jn.00263.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Wolpaw JR, Lee CL. Memory traces in primate spinal cord produced by operant conditioning of H-reflex. J Neurophysiol 1989; 61: 563–572. doi: 10.1152/jn.1989.61.3.563 [DOI] [PubMed] [Google Scholar]
  • 34.Funase K, Miles TS. Observations on the variability of the H reflex in human soleus. Muscle Nerve 1999; 22: 341–346. doi: [DOI] [PubMed] [Google Scholar]
  • 35.Herzfeld DJ, Shadmehr R. Motor variability is not noise, but grist for the learning mill. Nat Neurosci 2014; 17: 149–150. doi: 10.1038/nn.3633 [DOI] [PubMed] [Google Scholar]
  • 36.Wolpert DM, Diedrichsen J, Flanagan JR. Principles of sensorimotor learning. Nat Rev Neurosci 2011; 12: 739–751. doi: 10.1038/nrn3112 [DOI] [PubMed] [Google Scholar]
  • 37.Wolpert DM, Flanagan JR. Computations underlying sensorimotor learning. Curr Opin Neurobiol 2016; 37: 7–11. doi: 10.1016/j.conb.2015.12.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Nikooyan AA, Ahmed AA. Reward feedback accelerates motor learning. J Neurophysiol 2015; 113: 633–646. doi: 10.1152/jn.00032.2014 [DOI] [PubMed] [Google Scholar]
  • 39.Holland P, Codol O, Galea JM. Contribution of explicit processes to reinforcement-based motor learning. J Neurophysiol 2018; 119: 2241–2255. doi: 10.1152/jn.00901.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Guadagnoli MA, Lee TD. Challenge Point: A Framework for Conceptualizing the Effects of Various Practice Conditions in Motor Learning. J Mot Behav 2004; 36: 212–224. doi: 10.3200/JMBR.36.2.212-224 [DOI] [PubMed] [Google Scholar]
  • 41.Schmidt RA. Frequent Augmented Feedback Can Degrade Learning: Evidence and Interpretations. In: Requin J, Stelmach GE (eds) Tutorials in Motor Neuroscience. Dordrecht: Springer Netherlands, pp. 59–75. [Google Scholar]
  • 42.Schmidt RA, Walter CB. Knowledge of Results and Motor Learning: A Review and Critical Reappraisal. Psychol Bull 1984; 95: 355–386. [PubMed] [Google Scholar]
  • 43.Greenwood RJ, McMillan TM, Barnes MP, et al. Handbook of Neurological Rehabilitation. Psychology Press, 2005. [Google Scholar]
  • 44.Ramot M, Grossman S, Friedman D, et al. Covert neurofeedback without awareness shapes cortical network spontaneous connectivity. Proc Natl Acad Sci 2016; 113: E2413–E2420. doi: 10.1073/pnas.1516857113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Cortese A, Amano K, Koizumi A, et al. Multivoxel neurofeedback selectively modulates confidence without changing perceptual performance. Nat Commun 2016; 7: 13669. doi: 10.1038/ncomms13669 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Thompson AK, Pomerantz FR, Wolpaw JR. Operant Conditioning of a Spinal Reflex Can Improve Locomotion after Spinal Cord Injury in Humans. J Neurosci 2013; 33: 2365–2375. doi: 10.1523/JNEUROSCI.3968-12.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Rane C, Tyagi K, Manry M. Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions. Epub ahead of print 10 August 2023. doi: 10.48550/arXiv.2308.05724 [DOI] [Google Scholar]
  • 48.Granato G, Cartoni E, Rold FD, et al. Integrating unsupervised and reinforcement learning in human categorical perception: A computational model. PLOS ONE 2022; 17: e0267838. doi: 10.1371/journal.pone.0267838 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Brinkworth RSA, Tuncer M, Tucker KJ, et al. Standardization of H-reflex analyses. J Neurosci Methods 2007; 162: 1–7. doi: 10.1016/j.jneumeth.2006.11.020 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Umer Asgher

Transfer Alert

This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present.

25 Sep 2023

PONE-D-23-20613OPERANT REFLEX CONDITIONING SIMULATION ENVIRONMENT REVEALS EFFECTS OF FEEDBACK PARAMETERSPLOS ONE

Dear Dr. Sulzer,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

Based on the reviewers comments and feedback, "major revision is recommended and re-submit". For detailed comments to Authors, please refer reviewers' feedback section (appended below).

==============================

Please submit your revised manuscript by Nov 09 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Umer Asgher, PhD

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match.

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

3. Please update your submission to use the PLOS LaTeX template. The template and more information on our requirements for LaTeX submissions can be found at http://journals.plos.org/plosone/s/latex.

4. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

5. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well.

6. We note that Figures 1 and 3 in your submission contain copyrighted images. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For more information, see our copyright guidelines: http://journals.plos.org/plosone/s/licenses-and-copyright.

We require you to either (1) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (2) remove the figures from your submission:

1. You may seek permission from the original copyright holder of Figures 1 and 3 to publish the content specifically under the CC BY 4.0 license.

We recommend that you contact the original copyright holder with the Content Permission Form (http://journals.plos.org/plosone/s/file?id=7c09/content-permission-form.pdf) and the following text:

“I request permission for the open-access journal PLOS ONE to publish XXX under the Creative Commons Attribution License (CCAL) CC BY 4.0 (http://creativecommons.org/licenses/by/4.0/). Please be aware that this license allows unrestricted use and distribution, even commercially, by third parties. Please reply and provide explicit written permission to publish XXX under a CC BY license and complete the attached form.”

Please upload the completed Content Permission Form or other proof of granted permissions as an "Other" file with your submission.

In the figure caption of the copyrighted figure, please include the following text: “Reprinted from [ref] under a CC BY license, with permission from [name of publisher], original copyright [original copyright year].”

2. If you are unable to obtain permission from the original copyright holder to publish these figures under the CC BY 4.0 license or if the copyright holder’s requirements are incompatible with the CC BY 4.0 license, please either i) remove the figure or ii) supply a replacement figure that complies with the CC BY 4.0 license. Please check copyright information on all replacement figures and update the figure caption with source information. If applicable, please specify in the figure caption text when a figure is similar but not identical to the original image and is therefore for illustrative purposes only.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I will be evaluating this research based on PLOS ONE predefined criteria. Below is my review for “OPERANT REFLEX CONDITIONING SIMULATION ENVIRONMENT REVEALS EFFECTS OF FEEDBACK PARAMETERS”.

1. The study presents the results of original research.

Yes. To my knowledge there is no prior research that demonstrates the impact of feedback type, signal quality and success threshold on performance and strategy of a visuomotor task. However, there are several studies that have looked at these types of tasks and the strategy used across learner types within both implicit and explicit domains. It would improve the scholarship and emphasis on originality of this research by including the following references.

Brooks, V., Hipperath, F., Brooks, M., Ross, H., & Freund, H. (1995). Learning What and How in a Human Motor Task. Learning & Memory, 2(5), 225–243. https://doi.org/10.1101/lm.2.5.225

Hooyman, A., Gordon, J., & Winstein, C. (2021). Unique behavioral strategies in visuomotor learning: Hope for the non-learner. Human Movement Science, 79, 102858. https://doi.org/10.1016/j.humov.2021.102858

Holland, P., Codol, O., & Galea, J. M. (2018). Contribution of explicit processes to reinforcement-based motor learning. Journal of Neurophysiology, 119(6), 2241–2255. https://doi.org/10.1152/jn.00901.2017

Each paper identified “non-learners” and how the use, or lack thereof, of explicit and implicit strategies drives performance.

2. Results reported have not been published elsewhere.

I believe this to be correct, with the exception of a preprint in bioRxiv.

3. Experiments, statistics, and other analyses are performed to a high technical standard and are described in sufficient detail.

I am uncertain of how the learning of this task is equivalent to learning how to modulate actual H-reflex. Is there prior evidence that shows how learning capability of this task is related to individual capability to change H-reflex? Clarification on how learning of this task is related to learning of H-reflex is needed. What is lost if the simulated H-reflex is removed and every step after decision gain in figure 1 is maintained. What difference in visual feedback would this create that requires the simulated H-reflex to be included? I do recognize that you compare the virtual performance here to that of real performance of H-reflex conditioning from previously collected data [23]. However, this would at best represent an association between these types of learning and not necessarily a transfer of learning.

What is the y axis of figure 7? In the mixed effect model did you also include trial number as a fixed effect? Is the outcome variable for the trained LMM the same as hsim? Do you train an LMM on the virtual reflex data and then test it on the real data? I am having difficulty translating your stat method to figure 7. I thought hsim ranged from 0 to 1 but figure 7 shows an outcome variable ranging from 50 to 250? Also, for the fixed effect of variability, is that the biological variability or some other form of variability? Typically, in validation, although correlation is acceptable, other metrics of validation are MAE, MSE and RMSE. Additionally, you may consider replacing figure 7 with a Bland-Altman plot to demonstrate the agreement between real and simulated. Lastly, I think you need to provide some threshold by which the model is validated or not. To have a strong correlation in one out of 7 test participants isn’t strong evidence of validation. Alternatively, you could generate random data to be compared to the validated data and perform a contrast between the error in predictive accuracy.

Please provide a little more context for how to interpret hsim, the primary performance outcome measure. Please provide an example of what an hsim of 1 versus an hsim of 0 represents.

Could you please provide a visual of the Kr versus Kp versus KpKr display that participants would see under each condition?

Can you please include individual dots in the bar graphs to represent individual participant performance. It would also be good to use boxplots instead of bar graphs.

In the results please provide model estimates and confidence intervals of performance across conditions and interactions. You provide pairwise comparisons but condition performance should also be reported.

Although you demonstrate that mean group performance among the LV/KP condition had the worse performance and was the most aggressive this doesn’t really tell us if greater aggression/exploration is related to worse performance or if this is the case across all conditions. A similar situation appears in the reward threshold to variability result as well. IT would be helpful to see scatter plots of aggression versus performance stratified by group to better understand if these data support hypothesis 3: “the difficult reward criterion will worsen performance and make operant strategy less aggressive”.

4. Conclusions are presented in an appropriate fashion and are supported by the data.

I think the conclusions are presented well but I am not sure that results from this study indicate that explicit processes play a role in operant H-reflex conditioning, and this process is modulated by feedback parameters. I think the results the role of explicit processes on a visuomotor skill and how they are modified due to feedback parameters and feedback noise.

5. The article is presented in an intelligible fashion and is written in standard English.

Yes.

6. The research meets all applicable standards for the ethics of experimentation and research integrity.

Yes. This experiment examined visuomotor strategy among young, non-disabled adults. All participants knowingly gave consent to participate in this experiment.

7. The article adheres to appropriate reporting guidelines and community standards for data availability.

Yes. The data are available on open science framework.

Reviewer #2: Summary:

The manuscript discusses operant conditioning of neural activation and the role of feedback in shaping explicit decision-making processes. The authors developed a simulated operant conditioning environment and conducted an experiment involving 41 participants. They examined the impact of feedback type, signal quality, success threshold, and biological variability on operant conditioning performance and strategy. The study found that performance was influenced by variability, while operant strategy was affected by feedback type.

Review:

The manuscript explores an interesting and relevant topic in the field of operant conditioning and neural activation.

1. The introduction lacks a clear explanation of the significance of the research and its potential applications.

2. While the research aims to determine the explicit decision-making processes in response to feedback, the specific research questions or hypotheses are not explicitly stated in the introduction. It would be useful to outline the research objectives clearly.

3. The manuscript briefly describes the experimental setup but lacks sufficient detail about the web application game and the operant conditioning model. A more comprehensive description of the methodology, including the design of the game and how feedback was provided, would enhance the paper's clarity.

4. The paper mentions extracting parameters from real operant conditioning data but does not provide details about the statistical methods or analysis techniques used.

5. The authors should delve deeper into the significance of the observed relationships between feedback parameters and provide insights into how these findings can be applied or expanded upon. The manuscript should address the limitations of the study and suggest avenues for future research.

6. Consider including papers using dynamic activation in regards to neural networks. This will help make the manuscript take a more round shape regarding the literature review. Try to add some relevant papers from the PLOS Journal.

Here are a couple of references that might be helpful.

Rane, Chinmay, Kanishka Tyagi, and Michael Manry. "Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions." arXiv preprint arXiv:2308.05724 (2023).

Biswas, Koushik, Sandeep Kumar, Shilpak Banerjee, and Ashish Kumar Pandey. "TanhSoft—dynamic trainable activation functions for faster learning and better performance." IEEE Access 9 (2021): 120613-120623.

Karthikeyan, Anitha, Ashokkumar Srinivasan, Sundaram Arun, and Karthikeyan Rajagopal. "Complex network dynamics of a memristor neuron model with piecewise linear activation function." The European Physical Journal Special Topics 231, no. 22-23 (2022): 4089-4096.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Kanishka Tyagi

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Mar 21;19(3):e0300338. doi: 10.1371/journal.pone.0300338.r002

Author response to Decision Letter 0


15 Dec 2023

All responses to reviewers are included in the attached response to reviewers document.

Attachment

Submitted filename: Response to Reviewers_v2.docx

pone.0300338.s003.docx (500.4KB, docx)

Decision Letter 1

Manabu Sakakibara

8 Feb 2024

PONE-D-23-20613R1Simulated operant reflex conditioning environment reveals effects of feedback parametersPLOS ONE

Dear Dr. Sulzer,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

The original two referees have carefully reviewed the revised manuscript entitled, “Simulated operant reflex conditioning environment reveals effects of feedback parameters". Their comments are appended below. The second reviewer is satisfied the revision, while the first reviewer still has some minor concerns which should be considered before publication. This Academic Editor is sure the critical concerns make the manuscript strengthen. I will consider after receiving the revised manuscript with your replies to each comment.

Please submit your revised manuscript by Mar 24 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Manabu Sakakibara, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: I Don't Know

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: No

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you very much for the thorough and considerate response to my initial critique.

I only have a few minor recommendations before providing a decision of accept.

I am unsure if the data with your OSF directory are finalized but as I look at them in their current form they are nearly impossible to translate back to the experimental paradigm.

From what I can gather, each .txt file consists of the 10 conditions (which were presented in a random sequence?) with each condition consisting of 35 trials? I cannot tell which data point corresponds to which trial (or run) or condition and I believe that the three files for each participant represents the separate sessions? But then each column within each data file has no variable name. Even as I try to carefully read this manuscript, the data in the OSF only look like random numbers. Even when I try to plot them they just look like noise. These data either need a readme file and a complete reformatting to allow outside users to understand what they are looking at. Also, the data should not just be released as individual files. There should just be one large parent file with a column for participant ID, session, and whatever strtg means in the file name.

My motivation for this is because I wanted to look at the data as a response to my initial comment,

“In the mixed effect model did you also include trial number as a fixed effect?”

You responded that run number showed no effect on the model. This is surprising given this is a study on learning. There being no effect of run or trial number would indicate that participants do not improve overtime and thus no learning is actually taking place? When I look at fig 7, now with a y-axis label, I see now that their doesn’t appear to be any change in performance across the several runs, or at least it is impossible to infer from this graph given there is so much nested within each run (10 conditions presented at random, with 35 trials per condition). What is the explanation for this?

Reviewer #2: The author have diligently addressed all the changes and recommendations provided in the previous review. The revised paper is now fully prepared for submission. Thank you for taking all the constructive feedback and guidance.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Kanishka Tyagi

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Mar 21;19(3):e0300338. doi: 10.1371/journal.pone.0300338.r004

Author response to Decision Letter 1


21 Feb 2024

We thank the editor and reviewers for their valuable comments that have strengthened the manuscript. We have carefully reviewed these comments and we believe they have been sufficiently addressed. We have primarily addressed the comments in the manuscript but have included excerpts in this document where we felt necessary. Our responses are in red font.

The line numbers refer to lines in ‘Track Changes (Simple Markup)’ mode. Thank you very much for your time, effort, and attention.

Reviewer Comments:

Reviewer 1

1. I only have a few minor recommendations before providing a decision of accept.

I am unsure if the data with your OSF directory are finalized but as I look at them in their current form they are nearly impossible to translate back to the experimental paradigm.From what I can gather, each .txt file consists of the 10 conditions (which were presented in a random sequence?) with each condition consisting of 35 trials? I cannot tell which data point corresponds to which trial (or run) or condition and I believe that the three files for each participant represents the separate sessions? But then each column within each data file has no variable name. Even as I try to carefully read this manuscript, the data in the OSF only look like random numbers. Even when I try to plot them they just look like noise. These data either need a readme file and a complete reformatting to allow outside users to understand what they are looking at. Also, the data should not just be released as individual files. There should just be one large parent file with a column for participant ID, session, and whatever strtg means in the file name.

>> Thank you for the detailed and valuable comment. We have added a readme.txt file to the osf dataset that explains the data nomenclature.

2. My motivation for this is because I wanted to look at the data as a response to my initial comment,

“In the mixed effect model did you also include trial number as a fixed effect?”

You responded that run number showed no effect on the model. This is surprising given this is a study on learning. There being no effect of run or trial number would indicate that participants do not improve overtime and thus no learning is actually taking place? When I look at fig 7, now with a y-axis label, I see now that their doesn’t appear to be any change in performance across the several runs, or at least it is impossible to infer from this graph given there is so much nested within each run (10 conditions presented at random, with 35 trials per condition). What is the explanation for this?

>> Thank you for the valuable comment. We understand the confusion because in a typical motor learning paradigm there is run-by-run learning. However, this paradigm is an unskilled task, where no useful information is transferred to the next run. For instance, the target location randomly varies by run, and the variation of strategy to reach the target is not a difficult procedure, so any learning would show a ceiling effect quickly. We did not predict any learning between runs, and did not include the statistical analysis because it would be an unnecessary test. However, as could be inferred from Figure 7, there was no learning between runs based on an LMM we ran post-hoc. We have succinctly addressed the reviewer’s concern in the limitations paragraph:

Lines 503-6:

The current paradigm should not show run-by-run learning (Figure 7) because there was no strategy or information in this unskilled task to carry over to the following runs. However, future investigations could examine learning on a trial-by-trial basis [47, 48].

Attachment

Submitted filename: Response to Reviewers.docx

pone.0300338.s004.docx (20.9KB, docx)

Decision Letter 2

Manabu Sakakibara

27 Feb 2024

Simulated operant reflex conditioning environment reveals effects of feedback parameters

PONE-D-23-20613R2

Dear Dr. Sulzer,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Manabu Sakakibara, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you very much for conscientiously responding to each of prior comments. I hope this article can have a meaningful impact on the field.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

Acceptance letter

Manabu Sakakibara

12 Mar 2024

PONE-D-23-20613R2

PLOS ONE

Dear Dr. Sulzer,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Manabu Sakakibara

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. Effect of biological variability and feedback type on performance and operant strategy (moderate threshold).

    Values represent overall mean ± SE. Tukey HSD was used for pair-wise comparisons (↔) between conditions of high and low variability (HV, LV) and different feedback types (Kp, Kr, KpKr). Statistical significance (* p<0.05, **p<0.01, ***p<0.001, ****p<0.0001).

    (PDF)

    pone.0300338.s001.pdf (75.5KB, pdf)
    S2 Table. Effect of biological variability and reward threshold on performance and operant strategy.

    Values represent overall mean ± SE. Tukey HSD was used for pair-wise comparisons (↔) between conditions of high and low variability (HV, LV) and difficult, moderate, and easy threshold (DT, MT, ET). Statistical significance (* p<0.05, **p<0.01, ***p<0.001, ****p<0.0001).

    (PDF)

    pone.0300338.s002.pdf (91.6KB, pdf)
    Attachment

    Submitted filename: Response to Reviewers_v2.docx

    pone.0300338.s003.docx (500.4KB, docx)
    Attachment

    Submitted filename: Response to Reviewers.docx

    pone.0300338.s004.docx (20.9KB, docx)

    Data Availability Statement

    Data is available at: Sulzer, James, and Kyoungsoon Kim. 2023. “Operant Conditioning Simulation Paradigm.” OSF. May 26. DOI 10.17605/OSF.IO/Y8P47 https://osf.io/y8p47/.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES