Best practice: many, many files - in Oxygen, in filesystem in database
Posted: Mon Jun 03, 2024 12:23 pm
Dear all,
my question to everyone working with oxygen and xml is, how to work with a large amount of files. I mean large in number, several thousand.
For example, when you write a transformation, which converts a large 1 GB data-file into 200.000 seperate data-files.
There are several obstacles in the filesystem and in applications (editors, git, databases), which make it difficult to work with a large number, e.g. writing may be very fast, deleting is very slow, listing in Oxygen is very very slow, jumping in oxygen from one opened file to another is very very slow.
What is best practice to handle that? I develop and test with small datasets. Large dirs should be outside of git and oxygen? In database only production-ready results? Distribute files into dir-trees?
Thanks for your hints!
my question to everyone working with oxygen and xml is, how to work with a large amount of files. I mean large in number, several thousand.
For example, when you write a transformation, which converts a large 1 GB data-file into 200.000 seperate data-files.
There are several obstacles in the filesystem and in applications (editors, git, databases), which make it difficult to work with a large number, e.g. writing may be very fast, deleting is very slow, listing in Oxygen is very very slow, jumping in oxygen from one opened file to another is very very slow.
What is best practice to handle that? I develop and test with small datasets. Large dirs should be outside of git and oxygen? In database only production-ready results? Distribute files into dir-trees?
Thanks for your hints!