Skip to main content
. Author manuscript; available in PMC: 2024 Jan 19.
Published in final edited form as: Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2023 Aug 22;2023:227–236. doi: 10.1109/cvpr52729.2023.00030

Figure 2. Our architecture on the left and unsupervised learning losses on the right.

Figure 2.

Shared encoder comprised of VGG16 and FPN encodes first and second images. Point features are sampled at the location of ordered contour points indicated by rainbow colors from red to purple. Point features are inputted as query or key and value to the cross attentions. Lastly, shared FCNN takes the fused features and regresses forward Ott+1 or backward Ot+1t offsets. The cycle consistency, mechanical-normal, and mechanical-linear losses are shown in red color.