Skip to content
This repository was archived by the owner on Oct 14, 2025. It is now read-only.
This repository was archived by the owner on Oct 14, 2025. It is now read-only.

Memory issues when running parallel walkers #40

@FrederikHeber

Description

@FrederikHeber

There are some more issues when running things in parallel that I just stumbled upon. These are partly related to the new input pipeline, see #22.

  • cache() is a memory hog when batch_size is smaller than the dataset size. For example, full MNIST with batch_size of 10,000 will use more and more memory, while batch_size of 55,000 has constant memory after initialization. This might be related to tf 1.4.1 and might have improved with tensorflow newer versions.
  • Running multiple walkers (e.g., number_walkers of 2) on MNIST does not show the expected parallel load but fills the whole memory really quickly. This is independent of the above batch_size issue. In contrast to before, the memory filling stops sooner with smaller batch_size, e.g. 1,000. The DirectInputPipeline does not show this behavior. There, memory usage is roughly constant after initialization. In no case do we see parallel loading, there is just a single working thread.

Metadata

Metadata

Assignees

Labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions