Skip to main content
. 2020 Jul 9;10:11360. doi: 10.1038/s41598-020-68169-x

Figure 3.

Figure 3

Optimizing rate constants through backpropagation. Instead of choosing the rate constants manually, the BPTT algorithm was used to optimize them, along with the rest of the parameters of the network. (A) For two combinations of target rate constants (αs=0.684, αr=0.34, top panel) and (αs=0.34, αr=0.68, bottom panel) the optimization of the rate constants is plotted over the grid search results of Fig. 2. The target rate constants are marked by a green circle. Different initializations of the rate constants were tested, as indicated by the filled circles with different shades of green. The learned values after each epoch are marked by a cross of the same color. The dashed lines indicate the approximation where either α^s=1 or α^r=1. The intersection of both dashed lines indicates the Elman solution (blue circle). Independent of initialization, the learned rate constants all converge to the target rate constants, closely recovering the correct values. (B) The performance of the ARU model with learnable rate constants was compared with the classical Elman model. The ARU model performed significantly better than the Elman network (*p<1·10-11 and *p<1·10-6 for the top and bottom panels, respectively) for both data sets (error bars represent standard error over 20 repetitions).