-
Notifications
You must be signed in to change notification settings - Fork 246
/
convert_movie.py
111 lines (91 loc) · 4.5 KB
/
convert_movie.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
import mxnet as mx
import numpy as np
from sym import make_l1_sym
import argparse
import logging
import data
import cv2
def make_movie(model_path, ctx, source='test_idx', fname=None):
from parse import anaglyph, sbs
from parse import split
logging.basicConfig(level=logging.DEBUG)
n_batch = 50
batch_size = 10
start = 0 # 50*24*60/batch_size + 1
scale = (-15, 17)
data_frames = 1
flow_frames = 0
upsample = 4
data_shape = (384, 160)
udata_shape = (data_shape[0]*upsample, data_shape[1]*upsample)
base_shape = (432*upsample, 180*upsample)
method = 'multi2'
pos = [(24,10), (0,0), (0,20), (48,0), (48,20)]
norm = np.zeros((3, base_shape[1], base_shape[0]), dtype=np.float32)
boarder = 32*upsample
feather = np.ones((udata_shape[1], udata_shape[0]), dtype=np.float32)
for i in range(udata_shape[1]):
for j in range(udata_shape[0]):
feather[i, j] = min((min(i, j)+1.0)/boarder, 1.0)
for p in pos:
up = (p[0]*upsample, p[1]*upsample)
norm[:, up[1]:up[1]+udata_shape[1], up[0]:up[0]+udata_shape[0]] += feather
metric_xy = mx.metric.MAE()
metric_py = mx.metric.MAE()
test_data = data.Mov3dStack('data/lmdb', data_shape, batch_size, scale, output_depth=False,
data_frames=data_frames, flow_frames=flow_frames,
test_mode=True, source=source, upsample=upsample, base_shape=base_shape)
sym = make_l1_sym(scale, 'sum', data_frames, flow_frames, False, method=method, upsample=upsample)
init = mx.init.Load(model_path, verbose=True)
model = mx.model.FeedForward(ctx=ctx, symbol=sym, initializer=init)
model._init_params(dict(test_data.provide_data+test_data.provide_label))
lcap = cv2.VideoWriter()
pcap_ana = cv2.VideoWriter()
pcap_sbs = cv2.VideoWriter()
ycap_ana = cv2.VideoWriter()
ycap_sbs = cv2.VideoWriter()
lcap.open(fname+'_l.mkv', cv2.VideoWriter_fourcc(*'X264'), 24, base_shape)
pcap_ana.open(fname+'_ana_p.mkv', cv2.VideoWriter_fourcc(*'X264'), 24, base_shape)
pcap_sbs.open(fname+'_sbs_p.mkv', cv2.VideoWriter_fourcc(*'X264'), 24, base_shape)
data_names = [x[0] for x in test_data.provide_data]
model._init_predictor(test_data.provide_data)
data_arrays = [model._pred_exec.arg_dict[name] for name in data_names]
for i in range(n_batch):
X = np.zeros((batch_size,)+norm.shape, dtype=np.float32)
Y = np.zeros((batch_size,)+norm.shape, dtype=np.float32)
P = np.zeros((batch_size,)+norm.shape, dtype=np.float32)
for p in pos:
test_data.seek(start+i)
test_data.fix_p = p
batch = test_data.next()
mx.executor._load_data(batch, data_arrays)
model._pred_exec.forward(is_train=False)
up = (p[0]*upsample, p[1]*upsample)
X[:, :, up[1]:up[1]+udata_shape[1], up[0]:up[0]+udata_shape[0]] += batch.data[-1].asnumpy() * feather
Y[:, :, up[1]:up[1]+udata_shape[1], up[0]:up[0]+udata_shape[0]] += batch.label[0].asnumpy() * feather
P[:, :, up[1]:up[1]+udata_shape[1], up[0]:up[0]+udata_shape[0]] += model._pred_exec.outputs[0].asnumpy() * feather
X = np.clip(X/norm, 0, 255)
Y = np.clip(Y/norm, 0, 255)
P = np.clip(P/norm, 0, 255)
metric_py.update([mx.nd.array(Y)], [mx.nd.array(P)])
metric_xy.update([mx.nd.array(Y)], [mx.nd.array(X)])
print i, metric_py.get(), metric_xy.get()
X = X.astype(np.uint8).transpose((0, 2, 3, 1))
Y = Y.astype(np.uint8).transpose((0, 2, 3, 1))
P = P.astype(np.uint8).transpose((0, 2, 3, 1))
for i in range(X.shape[0]):
lcap.write(X[i])
for i in range(Y.shape[0]):
ycap_ana.write(anaglyph(X[i], Y[i]))
ycap_sbs.write(sbs(X[i], Y[i]))
for i in range(P.shape[0]):
pcap_ana.write(anaglyph(X[i], P[i]))
pcap_sbs.write(sbs(X[i], P[i]))
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='Convert movie to 3D')
parser.add_argument('model_path', type=str, default='deep3d-0050.params', help='Path to model parameters file')
parser.add_argument('--ctx', type=int, default=0, help='GPU id to run on')
parser.add_argument('--source', type=str, default='test_idx', help='Source index prefix')
parser.add_argument('--output', type=str, default='output', help='Output prefix')
args = parser.parse_args()
make_movie(args.model_path, mx.gpu(args.ctx), args.source, args.output)