I used the hyperparameter you gave me, but I failed to reproduce the result on the paper.
After running eval.py, I only get an accuracy of about 95.66% .
In your paper, in this case 0.51 errors(%) (shown Table 5). is it means an accuracy of 99.5%?
Is there something I'm missing here?
I also refer to your response. (issues/3) .
I visuallized your log.
The accuracy here is 97.59%. It is different from the result of your paper.( 0.51 errors(%) (shown Table 5))
Any thoughts, or exact instructions on how to replicate any of the results from the paper, would be greatly appreciated.
<SVHN -> MNIST>
flags of train.py
"target_dataset": "mnist3",
"walker_weight_envelope_delay": "500",
"max_checkpoints": 5,
"new_size": 32,
"dataset": "svhn",
"sup_per_batch": 100,
"decay_steps": 9000,
"unsup_batch_size": 1000,
"sup_per_class": -1,
"walker_weight_envelope_steps": 1,
"walker_weight_envelope": "linear",
"visit_weight_envelope": "linear",
"architecture": "svhn_model",
"visit_weight": 0.2,
"max_steps": "12000"
flags of eval.py
flags.DEFINE_string('dataset', 'mnist3', 'Which dataset to work on.')
flags.DEFINE_string('architecture', 'svhn_model', 'Which dataset to work on.')
flags.DEFINE_integer('eval_batch_size', 500, 'Batch size for eval loop.')
flags.DEFINE_integer('new_size', 32, 'If > 0, resize image to this width/height.'
'Needs to match size used for training.')
flags.DEFINE_integer('emb_size', 128,
'Size of the embeddings to learn.')
flags.DEFINE_integer('eval_interval_secs', 300,
'How many seconds between executions of the eval loop.')
flags.DEFINE_string('logdir', '/storage/transfer_learning/log2/semisup',
'Where the checkpoints are stored '
'and eval events will be written to.')
flags.DEFINE_string('master', '',
'BNS name of the TensorFlow master to use.')
flags.DEFINE_integer('timeout', 1200,
'The maximum amount of time to wait between checkpoints. '
'If left as None
, then the process will wait '
'indefinitely.')