import torch import numpy as np import torch.nn as nn import torch.nn.functional as F from model.warplayer import warp device = torch.device("cuda" if torch.cuda.is_available() else "cpu") def conv_wo_act(in_planes, out_planes, kernel_size=3, stride=1, padding=1, dilation=1): return nn.Sequential( nn.Conv2d(in_planes, out_planes, kernel_size=kernel_size, stride=stride, padding=padding, dilation=dilation, bias=False), nn.BatchNorm2d(out_planes), ) def conv(in_planes, out_planes, kernel_size=3, stride=1, padding=1, dilation=1): return nn.Sequential( nn.Conv2d(in_planes, out_planes, kernel_size=kernel_size, stride=stride, padding=padding, dilation=dilation, bias=False), nn.BatchNorm2d(out_planes), nn.PReLU(out_planes) ) class ResBlock(nn.Module): def __init__(self, in_planes, out_planes, stride=1): super(ResBlock, self).__init__() if in_planes == out_planes and stride == 1: self.conv0 = nn.Identity() else: self.conv0 = nn.Conv2d(in_planes, out_planes, 3, stride, 1, bias=False) self.conv1 = conv(in_planes, out_planes, 5, stride, 2) self.conv2 = conv_wo_act(out_planes, out_planes, 3, 1, 1) self.relu1 = nn.PReLU(1) self.relu2 = nn.PReLU(out_planes) self.fc1 = nn.Conv2d(out_planes, 16, kernel_size=1, bias=False) self.fc2 = nn.Conv2d(16, out_planes, kernel_size=1, bias=False) def forward(self, x): y = self.conv0(x) x = self.conv1(x) x = self.conv2(x) w = x.mean(3, True).mean(2, True) w = self.relu1(self.fc1(w)) w = torch.sigmoid(self.fc2(w)) x = self.relu2(x * w + y) return x class IFBlock(nn.Module): def __init__(self, in_planes, scale=1, c=64): super(IFBlock, self).__init__() self.scale = scale self.conv0 = conv(in_planes, c, 5, 2, 2) self.res0 = ResBlock(c, c) self.res1 = ResBlock(c, c) self.res2 = ResBlock(c, c) self.res3 = ResBlock(c, c) self.res4 = ResBlock(c, c) self.res5 = ResBlock(c, c) self.conv1 = nn.Conv2d(c, 8, 3, 1, 1) self.up = nn.PixelShuffle(2) def forward(self, x): if self.scale != 1: x = F.interpolate(x, scale_factor=1. / self.scale, mode="bilinear", align_corners=False) x = self.conv0(x) x = self.res0(x) x = self.res1(x) x = self.res2(x) x = self.res3(x) x = self.res4(x) x = self.res5(x) x = self.conv1(x) flow = self.up(x) if self.scale != 1: flow = F.interpolate(flow, scale_factor=self.scale, mode="bilinear", align_corners=False) return flow class IFNet(nn.Module): def __init__(self): super(IFNet, self).__init__() self.block0 = IFBlock(6, scale=8, c=192) self.block1 = IFBlock(8, scale=4, c=128) self.block2 = IFBlock(8, scale=2, c=96) self.block3 = IFBlock(8, scale=1, c=48) def forward(self, x, scale=1.0): x = F.interpolate(x, scale_factor=0.5 * scale, mode="bilinear", align_corners=False) flow0 = self.block0(x) F1 = flow0 warped_img0 = warp(x[:, :3], F1) warped_img1 = warp(x[:, 3:], -F1) flow1 = self.block1(torch.cat((warped_img0, warped_img1, F1), 1)) F2 = (flow0 + flow1) warped_img0 = warp(x[:, :3], F2) warped_img1 = warp(x[:, 3:], -F2) flow2 = self.block2(torch.cat((warped_img0, warped_img1, F2), 1)) F3 = (flow0 + flow1 + flow2) warped_img0 = warp(x[:, :3], F3) warped_img1 = warp(x[:, 3:], -F3) flow3 = self.block3(torch.cat((warped_img0, warped_img1, F3), 1)) F4 = (flow0 + flow1 + flow2 + flow3) F4 = F.interpolate(F4, scale_factor=1 / scale, mode="bilinear", align_corners=False) / scale return F4, [F1, F2, F3, F4] if __name__ == '__main__': img0 = torch.zeros(3, 3, 256, 256).float().to(device) img1 = torch.tensor(np.random.normal( 0, 1, (3, 3, 256, 256))).float().to(device) imgs = torch.cat((img0, img1), 1) flownet = IFNet() flow, _ = flownet(imgs) print(flow.shape)