1414device  =  torch .device ('cuda'  if  torch .cuda .is_available () else  'cpu' )
1515
1616def  load_image (image_path , transform = None ):
17-     image  =  Image .open (image_path )
17+     image  =  Image .open (image_path ). convert ( 'RGB' ) 
1818    image  =  image .resize ([224 , 224 ], Image .LANCZOS )
1919
2020    if  transform  is  not None :
@@ -69,13 +69,13 @@ def main(args):
6969if  __name__  ==  '__main__' :
7070    parser  =  argparse .ArgumentParser ()
7171    parser .add_argument ('--image' , type = str , required = True , help = 'input image for generating caption' )
72-     parser .add_argument ('--encoder_path' , type = str , default = 'models/encoder-2-1000.ckpt ' , help = 'path for trained encoder' )
73-     parser .add_argument ('--decoder_path' , type = str , default = 'models/decoder-2-1000.ckpt ' , help = 'path for trained decoder' )
72+     parser .add_argument ('--encoder_path' , type = str , default = 'models/encoder-5-3000.pkl ' , help = 'path for trained encoder' )
73+     parser .add_argument ('--decoder_path' , type = str , default = 'models/decoder-5-3000.pkl ' , help = 'path for trained decoder' )
7474    parser .add_argument ('--vocab_path' , type = str , default = 'data/vocab.pkl' , help = 'path for vocabulary wrapper' )
7575
7676    # Model parameters (should be same as paramters in train.py) 
7777    parser .add_argument ('--embed_size' , type = int  , default = 256 , help = 'dimension of word embedding vectors' )
7878    parser .add_argument ('--hidden_size' , type = int  , default = 512 , help = 'dimension of lstm hidden states' )
7979    parser .add_argument ('--num_layers' , type = int  , default = 1 , help = 'number of layers in lstm' )
8080    args  =  parser .parse_args ()
81-     main (args )
81+     main (args )
0 commit comments