Tensorflow 张量流指数移动平均

Tensorflow 张量流指数移动平均,tensorflow,moving-average,Tensorflow,Moving Average,我不知道如何使tf.train.ExponentialMovingAverage工作。下面是在简单的y=x*w方程中查找w的简单代码m是移动平均线。为什么m的代码返回None?如何让它返回移动平均值 import numpy as np import tensorflow as tf w = tf.Variable(0, dtype=tf.float32) ema = tf.train.ExponentialMovingAverage(decay=0.9) m = ema.apply([w])

我不知道如何使
tf.train.ExponentialMovingAverage
工作。下面是在简单的
y=x*w
方程中查找
w
的简单代码
m
是移动平均线。为什么
m
的代码返回
None
?如何让它返回移动平均值

import numpy as np
import tensorflow as tf

w = tf.Variable(0, dtype=tf.float32)
ema = tf.train.ExponentialMovingAverage(decay=0.9)
m = ema.apply([w])

x = tf.placeholder(tf.float32, [None])
y = tf.placeholder(tf.float32, [None])
y_ = tf.multiply(x, w)

with tf.control_dependencies([m]):
    loss = tf.reduce_sum(tf.square(tf.subtract(y, y_)))
    optimizer = tf.train.GradientDescentOptimizer(learning_rate=0.001)
    train = optimizer.minimize(loss)

with tf.Session() as sess:
    sess.run(tf.global_variables_initializer())
    for i in range(20):
        _, w_, m_ = sess.run([train, w, m], feed_dict={x: [1], y: [10]})
        print(w_, ',', m_)
输出为:

0.02 , None
0.03996 , None
0.0598801 , None
0.0797603 , None
0.0996008 , None
0.119402 , None
0.139163 , None
0.158884 , None
0.178567 , None
0.19821 , None
0.217813 , None
0.237378 , None
0.256903 , None
0.276389 , None
0.295836 , None
0.315244 , None
0.334614 , None
0.353945 , None
0.373237 , None
0.39249 , None

这是因为
m
(python)变量不保存操作的结果,而是保存操作本身。见文件:

要访问平均值,需要在图形中创建新元素:

av = ema.average(w)
然后:

_, w_, av_ = sess.run([train, w, av], feed_dict={x: [1], y: [10]})
print(w_, ',', av_)
将打印

[0.020000001, 0.0]
[0.039960001, 0.0020000006]
[0.059880082, 0.0057960013]
[0.07976032, 0.01120441]
[0.099600799, 0.018060002]

完整代码
嗨,在这种情况下,梯度下降如何在变量
w
上进行?更新
w
时,是使用原始值还是移动平均后的值?你能给我一些建议吗?谢谢
[0.020000001, 0.0]
[0.039960001, 0.0020000006]
[0.059880082, 0.0057960013]
[0.07976032, 0.01120441]
[0.099600799, 0.018060002]
import tensorflow as tf

w = tf.Variable(0, dtype=tf.float32)
ema = tf.train.ExponentialMovingAverage(decay=0.9)
m = ema.apply([w])
av = ema.average(w)

x = tf.placeholder(tf.float32, [None])
y = tf.placeholder(tf.float32, [None])
y_ = tf.multiply(x, w)

with tf.control_dependencies([m]):
    loss = tf.reduce_sum(tf.square(tf.subtract(y, y_)))
    optimizer = tf.train.GradientDescentOptimizer(learning_rate=0.001)
    train = optimizer.minimize(loss)

with tf.Session() as sess:
    sess.run(tf.global_variables_initializer())
    for i in range(20):
        _, w_, av_ = sess.run([train, w, av], feed_dict={x: [1], y: [10]})
        print(w_, ',', av_)