How to do large matrix decomposition with GPU in TensorflowTensorFlow: How to measure how much GPU memory each tensor takes?In tensorflow, how does one access a scalar tensor value before it is moved to the GPU?DMA between CPU and GPU in TensorFlowCPU/GPU Memory Usage with TensorflowTensorFlow GPU memoryHow does TensorFlow use both shared and dedicated GPU memory on the GPU on Windows 10?tensorflow not using gpu - prime number programtf.device('CPU:0') still utilizes gpu memorywhy Tensorflow-gpu is still using cpu

Commencez à vous connecter -- I don't understand the phrasing of this

What's the reason for the decade jump in the recent X-Men trilogy?

Why did the AvroCar fail to fly above 3 feet?

Can Mage Hand be used to indirectly trigger an attack?

What is the theme of analysis?

ISP is not hashing the password I log in with online. Should I take any action?

Can Dive Down protect a creature against Pacifism?

Is it possible to install Firefox on Ubuntu with no desktop enviroment?

Why does this Apple //e drops into system monitor when booting?

What is the color associated with lukewarm?

What do you call the action of "describing events as they happen" like sports anchors do?

Harley Davidson clattering noise from engine, backfire and failure to start

Idiom for 'person who gets violent when drunk"

Dedicated bike GPS computer over smartphone

I sent an angry e-mail to my interviewers about a conflict at my home institution. Could this affect my application?

Why did the Death Eaters wait to reopen the Chamber of Secrets?

What publication claimed that Michael Jackson died in a nuclear holocaust?

Opposite of "Concerto Grosso"?

Is fission/fusion to iron the most efficient way to convert mass to energy?

I received a gift from my sister who just got back from

Parallelized for loop in Bash

Are athletes' college degrees discounted by employers and graduate school admissions?

Must a CPU have a GPU if the motherboard provides a display port (when there isn't any separate video card)?

Is it true that "only photographers care about noise"?



How to do large matrix decomposition with GPU in Tensorflow


TensorFlow: How to measure how much GPU memory each tensor takes?In tensorflow, how does one access a scalar tensor value before it is moved to the GPU?DMA between CPU and GPU in TensorFlowCPU/GPU Memory Usage with TensorflowTensorFlow GPU memoryHow does TensorFlow use both shared and dedicated GPU memory on the GPU on Windows 10?tensorflow not using gpu - prime number programtf.device('CPU:0') still utilizes gpu memorywhy Tensorflow-gpu is still using cpu






.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty height:90px;width:728px;box-sizing:border-box;








1















I am trying to do a matrix decomposition (or tucker decomposition on a tensor) in Tensorflow with GPU. I have tensorflow-gpu, my NVidia GPU has 4GB RAM. My problem is that my input matrix is huge, millions of rows and millions of columns and the size of the matrix is more than 5GB in memory. So each time Tensorflow gives me an out of memory (OOM) error. (If I turn off GPU, the whole process can run successfully in CPU using system RAM. Of course, the speed is slow.)



I did some research on Tensorflow and on NVidia CUDA lib. CUDA seems has a "unified memory" mechanism so the system RAM and GPU RAM share one address book. Yet no further details found.



I wonder if Tensorflow supports some memory sharing mechanism such that I can generate input in system RAM? (Since I want to use GPU to accelerate the calculations) And GPU can do the calculation piece by piece.










share|improve this question
























  • Can this decomposition be split into smaller fragments manually? That’s your best bet. Shared memory architectures usually starve the GPU and you lose a lot of the speed advantage.

    – Kuba Ober
    Mar 25 at 16:36











  • Thank you for your cut-to-the-point comments, Kuba. I haven't found any easy/efficient ways to do it. One reason is I didn't find an easy way to split this SVD-similar job. Secondly, I am not sure it's worth it considering the difference of different RAM's bandwidth.

    – Amartin
    Mar 26 at 16:45

















1















I am trying to do a matrix decomposition (or tucker decomposition on a tensor) in Tensorflow with GPU. I have tensorflow-gpu, my NVidia GPU has 4GB RAM. My problem is that my input matrix is huge, millions of rows and millions of columns and the size of the matrix is more than 5GB in memory. So each time Tensorflow gives me an out of memory (OOM) error. (If I turn off GPU, the whole process can run successfully in CPU using system RAM. Of course, the speed is slow.)



I did some research on Tensorflow and on NVidia CUDA lib. CUDA seems has a "unified memory" mechanism so the system RAM and GPU RAM share one address book. Yet no further details found.



I wonder if Tensorflow supports some memory sharing mechanism such that I can generate input in system RAM? (Since I want to use GPU to accelerate the calculations) And GPU can do the calculation piece by piece.










share|improve this question
























  • Can this decomposition be split into smaller fragments manually? That’s your best bet. Shared memory architectures usually starve the GPU and you lose a lot of the speed advantage.

    – Kuba Ober
    Mar 25 at 16:36











  • Thank you for your cut-to-the-point comments, Kuba. I haven't found any easy/efficient ways to do it. One reason is I didn't find an easy way to split this SVD-similar job. Secondly, I am not sure it's worth it considering the difference of different RAM's bandwidth.

    – Amartin
    Mar 26 at 16:45













1












1








1








I am trying to do a matrix decomposition (or tucker decomposition on a tensor) in Tensorflow with GPU. I have tensorflow-gpu, my NVidia GPU has 4GB RAM. My problem is that my input matrix is huge, millions of rows and millions of columns and the size of the matrix is more than 5GB in memory. So each time Tensorflow gives me an out of memory (OOM) error. (If I turn off GPU, the whole process can run successfully in CPU using system RAM. Of course, the speed is slow.)



I did some research on Tensorflow and on NVidia CUDA lib. CUDA seems has a "unified memory" mechanism so the system RAM and GPU RAM share one address book. Yet no further details found.



I wonder if Tensorflow supports some memory sharing mechanism such that I can generate input in system RAM? (Since I want to use GPU to accelerate the calculations) And GPU can do the calculation piece by piece.










share|improve this question
















I am trying to do a matrix decomposition (or tucker decomposition on a tensor) in Tensorflow with GPU. I have tensorflow-gpu, my NVidia GPU has 4GB RAM. My problem is that my input matrix is huge, millions of rows and millions of columns and the size of the matrix is more than 5GB in memory. So each time Tensorflow gives me an out of memory (OOM) error. (If I turn off GPU, the whole process can run successfully in CPU using system RAM. Of course, the speed is slow.)



I did some research on Tensorflow and on NVidia CUDA lib. CUDA seems has a "unified memory" mechanism so the system RAM and GPU RAM share one address book. Yet no further details found.



I wonder if Tensorflow supports some memory sharing mechanism such that I can generate input in system RAM? (Since I want to use GPU to accelerate the calculations) And GPU can do the calculation piece by piece.







tensorflow gpu






share|improve this question















share|improve this question













share|improve this question




share|improve this question








edited Mar 25 at 16:31









Miroslav Glamuzina

2,88521223




2,88521223










asked Mar 25 at 1:26









AmartinAmartin

62




62












  • Can this decomposition be split into smaller fragments manually? That’s your best bet. Shared memory architectures usually starve the GPU and you lose a lot of the speed advantage.

    – Kuba Ober
    Mar 25 at 16:36











  • Thank you for your cut-to-the-point comments, Kuba. I haven't found any easy/efficient ways to do it. One reason is I didn't find an easy way to split this SVD-similar job. Secondly, I am not sure it's worth it considering the difference of different RAM's bandwidth.

    – Amartin
    Mar 26 at 16:45

















  • Can this decomposition be split into smaller fragments manually? That’s your best bet. Shared memory architectures usually starve the GPU and you lose a lot of the speed advantage.

    – Kuba Ober
    Mar 25 at 16:36











  • Thank you for your cut-to-the-point comments, Kuba. I haven't found any easy/efficient ways to do it. One reason is I didn't find an easy way to split this SVD-similar job. Secondly, I am not sure it's worth it considering the difference of different RAM's bandwidth.

    – Amartin
    Mar 26 at 16:45
















Can this decomposition be split into smaller fragments manually? That’s your best bet. Shared memory architectures usually starve the GPU and you lose a lot of the speed advantage.

– Kuba Ober
Mar 25 at 16:36





Can this decomposition be split into smaller fragments manually? That’s your best bet. Shared memory architectures usually starve the GPU and you lose a lot of the speed advantage.

– Kuba Ober
Mar 25 at 16:36













Thank you for your cut-to-the-point comments, Kuba. I haven't found any easy/efficient ways to do it. One reason is I didn't find an easy way to split this SVD-similar job. Secondly, I am not sure it's worth it considering the difference of different RAM's bandwidth.

– Amartin
Mar 26 at 16:45





Thank you for your cut-to-the-point comments, Kuba. I haven't found any easy/efficient ways to do it. One reason is I didn't find an easy way to split this SVD-similar job. Secondly, I am not sure it's worth it considering the difference of different RAM's bandwidth.

– Amartin
Mar 26 at 16:45












0






active

oldest

votes












Your Answer






StackExchange.ifUsing("editor", function ()
StackExchange.using("externalEditor", function ()
StackExchange.using("snippets", function ()
StackExchange.snippets.init();
);
);
, "code-snippets");

StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "1"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);

else
createEditor();

);

function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);



);













draft saved

draft discarded


















StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f55330237%2fhow-to-do-large-matrix-decomposition-with-gpu-in-tensorflow%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown

























0






active

oldest

votes








0






active

oldest

votes









active

oldest

votes






active

oldest

votes















draft saved

draft discarded
















































Thanks for contributing an answer to Stack Overflow!


  • Please be sure to answer the question. Provide details and share your research!

But avoid


  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.

To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f55330237%2fhow-to-do-large-matrix-decomposition-with-gpu-in-tensorflow%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

Kamusi Yaliyomo Aina za kamusi | Muundo wa kamusi | Faida za kamusi | Dhima ya picha katika kamusi | Marejeo | Tazama pia | Viungo vya nje | UrambazajiKuhusu kamusiGo-SwahiliWiki-KamusiKamusi ya Kiswahili na Kiingerezakuihariri na kuongeza habari

SQL error code 1064 with creating Laravel foreign keysForeign key constraints: When to use ON UPDATE and ON DELETEDropping column with foreign key Laravel error: General error: 1025 Error on renameLaravel SQL Can't create tableLaravel Migration foreign key errorLaravel php artisan migrate:refresh giving a syntax errorSQLSTATE[42S01]: Base table or view already exists or Base table or view already exists: 1050 Tableerror in migrating laravel file to xampp serverSyntax error or access violation: 1064:syntax to use near 'unsigned not null, modelName varchar(191) not null, title varchar(191) not nLaravel cannot create new table field in mysqlLaravel 5.7:Last migration creates table but is not registered in the migration table

은진 송씨 목차 역사 본관 분파 인물 조선 왕실과의 인척 관계 집성촌 항렬자 인구 같이 보기 각주 둘러보기 메뉴은진 송씨세종실록 149권, 지리지 충청도 공주목 은진현