代码拉取完成,页面将自动刷新
tf.compat.v1.train.exponential_decay(
learning_rate,
global_step,
decay_steps,
decay_rate,
staircase=False,
name=None
) -> Tensor
For more information, see tf.compat.v1.train.exponential_decay.
mindspore.nn.exponential_decay_lr(
learning_rate,
decay_rate,
total_step,
step_per_epoch,
decay_epoch,
is_stair=False
) -> list[float]
For more information, see mindspore.nn.exponential_decay_lr.
TensorFlow: calculate the learning rate based on the exponential decay function.
MindSpore: MindSpore API basically implements the same function as TensorFlow.
Categories | Subcategories | TensorFlow | MindSpore | Differences |
---|---|---|---|---|
Parameters | Parameter 1 | learning_rate | learning_rate | - |
Parameter 2 | global_step | total_step | Same function, different parameter names | |
Parameter 3 | decay_steps | decay_epoch | Same function, different parameter names | |
Parameter 4 | decay_rate | decay_rate | - | |
Parameter 5 | staircase | is_stair | Same function, different parameter names | |
Parameter 6 | name | - | Not involved | |
Parameter 7 | - | step_per_epoch | The number of steps per epoch, TensorFlow does not have this parameter |
The two APIs achieve the same function and have the same usage.
# TensorFlow
import tensorflow as tf
learning_rate = 1.0
decay_rate = 0.9
step_per_epoch = 2
epochs = 3
lr = []
for epoch in range(epochs):
learning_rate = tf.compat.v1.train.exponential_decay(learning_rate, epoch, step_per_epoch, decay_rate, staircase=True)
learning_rate = learning_rate().numpy().item()
lr.append(round(float(learning_rate), 2))
print(lr)
# [1.0, 1.0, 0.9]
# MindSpore
import mindspore.nn as nn
learning_rate = 1.0
decay_rate = 0.9
total_step = 3
step_per_epoch = 2
decay_epoch = 1
output = nn.exponential_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch)
print(output)
# [1.0, 1.0, 0.9]
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。