/aosp_15_r20/external/tensorflow/tensorflow/python/keras/optimizer_v2/ |
H A D | legacy_learning_rate_decay.py | 29 def exponential_decay(learning_rate, argument 183 def polynomial_decay(learning_rate, argument 284 def natural_exp_decay(learning_rate, argument 372 def inverse_time_decay(learning_rate, argument 455 def cosine_decay(learning_rate, global_step, decay_steps, alpha=0.0, name=None): argument 518 def cosine_decay_restarts(learning_rate, argument 595 def linear_cosine_decay(learning_rate, argument 680 def noisy_linear_cosine_decay(learning_rate, argument
|
H A D | adam.py | 106 learning_rate=0.001, argument 330 learning_rate=0.001, argument
|
H A D | adamax.py | 90 learning_rate=0.001, argument
|
H A D | ftrl.py | 103 learning_rate=0.001, argument
|
H A D | adadelta.py | 71 learning_rate=0.001, argument
|
H A D | nadam.py | 65 learning_rate=0.001, argument
|
H A D | gradient_descent.py | 101 learning_rate=0.01, argument
|
H A D | rmsprop.py | 96 learning_rate=0.001, argument
|
H A D | adagrad.py | 68 learning_rate=0.001, argument
|
/aosp_15_r20/external/tensorflow/tensorflow/python/training/ |
H A D | adagrad_test.py | 50 learning_rate = lambda: 3.0 function 314 learning_rate = lambda: 3.0 function
|
H A D | adam_test.py | 180 learning_rate = lambda: 0.001 function 398 learning_rate = lambda: 0.001 function
|
H A D | gradient_descent.py | 31 def __init__(self, learning_rate, use_locking=False, name="GradientDescent"): argument
|
H A D | proximal_gradient_descent.py | 37 def __init__(self, learning_rate, l1_regularization_strength=0.0, argument
|
H A D | momentum.py | 116 def __init__(self, learning_rate, momentum, argument
|
H A D | adadelta.py | 106 def __init__(self, learning_rate=0.001, rho=0.95, epsilon=1e-8, argument
|
H A D | proximal_adagrad.py | 39 def __init__(self, learning_rate, initial_accumulator_value=0.1, argument
|
H A D | rmsprop.py | 137 learning_rate, argument
|
H A D | adagrad_da.py | 45 learning_rate, argument
|
H A D | adagrad.py | 109 def __init__(self, learning_rate, initial_accumulator_value=0.1, argument
|
/aosp_15_r20/external/tensorflow/tensorflow/core/grappler/optimizers/ |
H A D | auto_parallel_test.cc | 41 Output learning_rate = ops::Const(s.WithOpName("learning_rate"), 0.01f, {1}); in TEST_F() local 137 Output learning_rate = ops::Const(s.WithOpName("learning_rate"), 0.01f, {1}); in TEST_F() local
|
/aosp_15_r20/external/pytorch/test/cpp/jit/ |
H A D | test_lite_trainer.cpp | 31 double learning_rate = 0.1, momentum = 0.1; in TEST() local 262 double learning_rate = 0.1, momentum = 0.1; in TEST() local
|
/aosp_15_r20/external/tensorflow/tensorflow/core/protobuf/tpu/ |
H A D | optimization_parameters.proto | 73 oneof learning_rate { oneof 490 LearningRate learning_rate = 13; field
|
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/mixed_precision/ |
H A D | loss_scale_optimizer.py | 848 def learning_rate(self): member in LossScaleOptimizer 852 def learning_rate(self, value): member in LossScaleOptimizer
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/mlir/tfr/examples/mnist/ |
H A D | mnist_train.py | 37 learning_rate = 0.001 variable
|
/aosp_15_r20/external/webrtc/rtc_base/ |
H A D | rolling_accumulator.h | 111 double ComputeWeightedMean(double learning_rate) const { in ComputeWeightedMean()
|