You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Oct 14, 2025. It is now read-only.
There are some more issues when running things in parallel that I just stumbled upon. These are partly related to the new input pipeline, see #22.
cache() is a memory hog when batch_size is smaller than the dataset size. For example, full MNIST with batch_size of 10,000 will use more and more memory, while batch_size of 55,000 has constant memory after initialization. This might be related to tf 1.4.1 and might have improved with tensorflow newer versions.
Running multiple walkers (e.g., number_walkers of 2) on MNIST does not show the expected parallel load but fills the whole memory really quickly. This is independent of the above batch_size issue. In contrast to before, the memory filling stops sooner with smaller batch_size, e.g. 1,000. The DirectInputPipeline does not show this behavior. There, memory usage is roughly constant after initialization. In no case do we see parallel loading, there is just a single working thread.