Home | History | Annotate | Download | only in v1
      1 path: "tensorflow.train.AdadeltaOptimizer"
      2 tf_class {
      3   is_instance: "<class \'tensorflow.python.training.adadelta.AdadeltaOptimizer\'>"
      4   is_instance: "<class \'tensorflow.python.training.optimizer.Optimizer\'>"
      5   is_instance: "<class \'tensorflow.python.training.tracking.base.Trackable\'>"
      6   is_instance: "<type \'object\'>"
      7   member {
      8     name: "GATE_GRAPH"
      9     mtype: "<type \'int\'>"
     10   }
     11   member {
     12     name: "GATE_NONE"
     13     mtype: "<type \'int\'>"
     14   }
     15   member {
     16     name: "GATE_OP"
     17     mtype: "<type \'int\'>"
     18   }
     19   member_method {
     20     name: "__init__"
     21     argspec: "args=[\'self\', \'learning_rate\', \'rho\', \'epsilon\', \'use_locking\', \'name\'], varargs=None, keywords=None, defaults=[\'0.001\', \'0.95\', \'1e-08\', \'False\', \'Adadelta\'], "
     22   }
     23   member_method {
     24     name: "apply_gradients"
     25     argspec: "args=[\'self\', \'grads_and_vars\', \'global_step\', \'name\'], varargs=None, keywords=None, defaults=[\'None\', \'None\'], "
     26   }
     27   member_method {
     28     name: "compute_gradients"
     29     argspec: "args=[\'self\', \'loss\', \'var_list\', \'gate_gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'grad_loss\'], varargs=None, keywords=None, defaults=[\'None\', \'1\', \'None\', \'False\', \'None\'], "
     30   }
     31   member_method {
     32     name: "get_name"
     33     argspec: "args=[\'self\'], varargs=None, keywords=None, defaults=None"
     34   }
     35   member_method {
     36     name: "get_slot"
     37     argspec: "args=[\'self\', \'var\', \'name\'], varargs=None, keywords=None, defaults=None"
     38   }
     39   member_method {
     40     name: "get_slot_names"
     41     argspec: "args=[\'self\'], varargs=None, keywords=None, defaults=None"
     42   }
     43   member_method {
     44     name: "minimize"
     45     argspec: "args=[\'self\', \'loss\', \'global_step\', \'var_list\', \'gate_gradients\', \'aggregation_method\', \'colocate_gradients_with_ops\', \'name\', \'grad_loss\'], varargs=None, keywords=None, defaults=[\'None\', \'None\', \'1\', \'None\', \'False\', \'None\', \'None\'], "
     46   }
     47   member_method {
     48     name: "variables"
     49     argspec: "args=[\'self\'], varargs=None, keywords=None, defaults=None"
     50   }
     51 }
     52