From 4ca6b1610f9fa65f8bd7d7c15059bfde18a2f02a Mon Sep 17 00:00:00 2001 From: Boris Fomitchev Date: Tue, 8 May 2018 00:56:35 -0700 Subject: Added data size and ONNX export options, FP16 inference is working --- models/models.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) (limited to 'models/models.py') diff --git a/models/models.py b/models/models.py index 0ba442f..805696f 100755 --- a/models/models.py +++ b/models/models.py @@ -10,7 +10,8 @@ def create_model(opt): from .ui_model import UIModel model = UIModel() model.initialize(opt) - print("model [%s] was created" % (model.name())) + if opt.verbose: + print("model [%s] was created" % (model.name())) if opt.isTrain and len(opt.gpu_ids): model = torch.nn.DataParallel(model, device_ids=opt.gpu_ids) -- cgit v1.2.3-70-g09d2