US 11,689,693 B2
Video frame interpolation method and device, computer readable storage medium
Yunhua Lu, Beijing (CN); Guannan Chen, Beijing (CN); Ran Duan, Beijing (CN); Lijie Zhang, Beijing (CN); and Hanwen Liu, Beijing (CN)
Assigned to BOE TECHNOLOGY GROUP CO., LTD., Beijing (CN)
Appl. No. 17/265,568
Filed by BOE TECHNOLOGY GROUP CO., LTD., Beijing (CN)
PCT Filed Apr. 30, 2020, PCT No. PCT/CN2020/088490
§ 371(c)(1), (2) Date Feb. 3, 2021,
PCT Pub. No. WO2021/217653, PCT Pub. Date Nov. 4, 2021.
Prior Publication US 2022/0116567 A1, Apr. 14, 2022
Int. Cl. H04N 7/01 (2006.01); G06N 20/00 (2019.01)
CPC H04N 7/0135 (2013.01) [G06N 20/00 (2019.01); H04N 7/0145 (2013.01)] 20 Claims
OG exemplary drawing
 
1. A video frame interpolation method, comprising:
inputting at least two image frames into a video frame interpolation model to obtain at least one frame-interpolation image frame, wherein the video frame interpolation model is obtained by training an initial model and the training the initial model comprises:
obtaining the initial model and sample images, and the training the initial model using a first loss to obtain a reference model;
obtaining a first training model comprising three identical reference models, selecting two first target sample image frames from the sample images according to a preset rule, and inputting the first target sample image frames into a first reference model of the first training model to obtain a first frame-interpolation result;
selecting two second target sample image frames from the sample images according to the preset rule, and inputting the second target sample image frames into a second reference model of the first training model to obtain a second frame-interpolation result;
selecting one frame-interpolation image from the first frame-interpolation result according to the preset rule and selecting one frame-interpolation image from the second frame-interpolation result according to the preset rule, wherein the frame-interpolation images selected are used as two third target sample image frames that are input into a third reference model of the first training model for calculation to obtain a third frame-interpolation result;
adjusting parameters of the first training model based on total loss of the first training model, wherein the total loss of the first training model is obtained based on the first frame-interpolation result, the second frame-interpolation result, and the third frame-interpolation result and the sample images, and the parameters are shared by each reference model of the first training model; and
using a parameter model obtained via a predetermined number of iterations as the video frame interpolation model.