How do you compute and/or apply the gradients of paticular parameters within a variable in TensorFlow?









up vote
0
down vote

favorite












I have four sets of TensorFlow Variables.



I am using these variables in an autoencoder based on the archtiecture found in this paper: http://users.cecs.anu.edu.au/~u5098633/papers/www15.pdf



Briefly, only the weights that are connected to certain input nodes are important for training. That is, if we have 10 input nodes, I only want to focus on, say, 5 of these nodes, and the weights that are connected to these nodes. My variable that models these connections is:



self.V = tf.Variable(tf.truncated_normal([input_size, hidden_size]), name='V') 


I am able to get these particular weights by using:



weights_to_update = tf.gather(self.V, [indices_of_weights]).



Now, to get the particular gradients of these weights, I have tried using



grads_and_vars = tf.Optimizer.compute_gradients(my_loss_function, weights_to_update)



which I thought would calculate the gradients of these particular weights, and ignore all other weights in the self.V variable. I would then use tf.Optimizer.apply_gradients(graqds_and_vars) to perform the update step.



However, the output I get from using compute_gradients with the var_list parameter as the weights I want is:



[(None, <tf.Tensor 'GatherV2:0' shape=(20, 200) dtype=float32>)] 


which shows that the shape is correct for what I want, but that this is None values. Subsequently, calling apply_gradients with what this returns leads to an error.



I have searched all over the internet to find out how to compute/apply gradients to a subset of a tf.Variable, but I cannot find out how to do this anywhere.



So, to sum up, for each training step, I would like to know how to train only a subset of parameters from each of my tf.Variable objects in my graph.










share|improve this question





















  • If your gradient is None, there is probably a problem with your network setup. Does simply using Optimizer.minimize(loss) work (i.e. run) and change all the weights?
    – cheersmate
    Nov 12 at 8:27














up vote
0
down vote

favorite












I have four sets of TensorFlow Variables.



I am using these variables in an autoencoder based on the archtiecture found in this paper: http://users.cecs.anu.edu.au/~u5098633/papers/www15.pdf



Briefly, only the weights that are connected to certain input nodes are important for training. That is, if we have 10 input nodes, I only want to focus on, say, 5 of these nodes, and the weights that are connected to these nodes. My variable that models these connections is:



self.V = tf.Variable(tf.truncated_normal([input_size, hidden_size]), name='V') 


I am able to get these particular weights by using:



weights_to_update = tf.gather(self.V, [indices_of_weights]).



Now, to get the particular gradients of these weights, I have tried using



grads_and_vars = tf.Optimizer.compute_gradients(my_loss_function, weights_to_update)



which I thought would calculate the gradients of these particular weights, and ignore all other weights in the self.V variable. I would then use tf.Optimizer.apply_gradients(graqds_and_vars) to perform the update step.



However, the output I get from using compute_gradients with the var_list parameter as the weights I want is:



[(None, <tf.Tensor 'GatherV2:0' shape=(20, 200) dtype=float32>)] 


which shows that the shape is correct for what I want, but that this is None values. Subsequently, calling apply_gradients with what this returns leads to an error.



I have searched all over the internet to find out how to compute/apply gradients to a subset of a tf.Variable, but I cannot find out how to do this anywhere.



So, to sum up, for each training step, I would like to know how to train only a subset of parameters from each of my tf.Variable objects in my graph.










share|improve this question





















  • If your gradient is None, there is probably a problem with your network setup. Does simply using Optimizer.minimize(loss) work (i.e. run) and change all the weights?
    – cheersmate
    Nov 12 at 8:27












up vote
0
down vote

favorite









up vote
0
down vote

favorite











I have four sets of TensorFlow Variables.



I am using these variables in an autoencoder based on the archtiecture found in this paper: http://users.cecs.anu.edu.au/~u5098633/papers/www15.pdf



Briefly, only the weights that are connected to certain input nodes are important for training. That is, if we have 10 input nodes, I only want to focus on, say, 5 of these nodes, and the weights that are connected to these nodes. My variable that models these connections is:



self.V = tf.Variable(tf.truncated_normal([input_size, hidden_size]), name='V') 


I am able to get these particular weights by using:



weights_to_update = tf.gather(self.V, [indices_of_weights]).



Now, to get the particular gradients of these weights, I have tried using



grads_and_vars = tf.Optimizer.compute_gradients(my_loss_function, weights_to_update)



which I thought would calculate the gradients of these particular weights, and ignore all other weights in the self.V variable. I would then use tf.Optimizer.apply_gradients(graqds_and_vars) to perform the update step.



However, the output I get from using compute_gradients with the var_list parameter as the weights I want is:



[(None, <tf.Tensor 'GatherV2:0' shape=(20, 200) dtype=float32>)] 


which shows that the shape is correct for what I want, but that this is None values. Subsequently, calling apply_gradients with what this returns leads to an error.



I have searched all over the internet to find out how to compute/apply gradients to a subset of a tf.Variable, but I cannot find out how to do this anywhere.



So, to sum up, for each training step, I would like to know how to train only a subset of parameters from each of my tf.Variable objects in my graph.










share|improve this question













I have four sets of TensorFlow Variables.



I am using these variables in an autoencoder based on the archtiecture found in this paper: http://users.cecs.anu.edu.au/~u5098633/papers/www15.pdf



Briefly, only the weights that are connected to certain input nodes are important for training. That is, if we have 10 input nodes, I only want to focus on, say, 5 of these nodes, and the weights that are connected to these nodes. My variable that models these connections is:



self.V = tf.Variable(tf.truncated_normal([input_size, hidden_size]), name='V') 


I am able to get these particular weights by using:



weights_to_update = tf.gather(self.V, [indices_of_weights]).



Now, to get the particular gradients of these weights, I have tried using



grads_and_vars = tf.Optimizer.compute_gradients(my_loss_function, weights_to_update)



which I thought would calculate the gradients of these particular weights, and ignore all other weights in the self.V variable. I would then use tf.Optimizer.apply_gradients(graqds_and_vars) to perform the update step.



However, the output I get from using compute_gradients with the var_list parameter as the weights I want is:



[(None, <tf.Tensor 'GatherV2:0' shape=(20, 200) dtype=float32>)] 


which shows that the shape is correct for what I want, but that this is None values. Subsequently, calling apply_gradients with what this returns leads to an error.



I have searched all over the internet to find out how to compute/apply gradients to a subset of a tf.Variable, but I cannot find out how to do this anywhere.



So, to sum up, for each training step, I would like to know how to train only a subset of parameters from each of my tf.Variable objects in my graph.







python tensorflow neural-network






share|improve this question













share|improve this question











share|improve this question




share|improve this question










asked Nov 11 at 1:03









Frederick

336




336











  • If your gradient is None, there is probably a problem with your network setup. Does simply using Optimizer.minimize(loss) work (i.e. run) and change all the weights?
    – cheersmate
    Nov 12 at 8:27
















  • If your gradient is None, there is probably a problem with your network setup. Does simply using Optimizer.minimize(loss) work (i.e. run) and change all the weights?
    – cheersmate
    Nov 12 at 8:27















If your gradient is None, there is probably a problem with your network setup. Does simply using Optimizer.minimize(loss) work (i.e. run) and change all the weights?
– cheersmate
Nov 12 at 8:27




If your gradient is None, there is probably a problem with your network setup. Does simply using Optimizer.minimize(loss) work (i.e. run) and change all the weights?
– cheersmate
Nov 12 at 8:27

















active

oldest

votes











Your Answer






StackExchange.ifUsing("editor", function ()
StackExchange.using("externalEditor", function ()
StackExchange.using("snippets", function ()
StackExchange.snippets.init();
);
);
, "code-snippets");

StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "1"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);

else
createEditor();

);

function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);



);













 

draft saved


draft discarded


















StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53244934%2fhow-do-you-compute-and-or-apply-the-gradients-of-paticular-parameters-within-a-v%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown






























active

oldest

votes













active

oldest

votes









active

oldest

votes






active

oldest

votes















 

draft saved


draft discarded















































 


draft saved


draft discarded














StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53244934%2fhow-do-you-compute-and-or-apply-the-gradients-of-paticular-parameters-within-a-v%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







這個網誌中的熱門文章

How to read a connectionString WITH PROVIDER in .NET Core?

In R, how to develop a multiplot heatmap.2 figure showing key labels successfully

Museum of Modern and Contemporary Art of Trento and Rovereto