I am in the process of writing an introduction to an article that will change the way in which a particular type of data is stored, allowing it to have more flexibility for particular tasks.

One argument that I have seen from reviewers in the past is that they do not see the need to keep additional data when no one is using it right now (citing that humans only make discernments to a particular level, so making models that can identify thousands of differences while maintaining the information that handles what is currently established is not necessary).  I would like to shed light on why it is important to be ahead in the data curve, being able to tackle the unforeseen problems that may arise in an area.  I feel like an article addressing this topic exists somewhere, but I cannot find the right search terms to produce it.

For context, I am not changing the amount of data that is stored, but rather the way in which it is stored.  Rather than to maintain just binary set intersections between an object and partition tiles of a space, I am maintain a topological relationship between the object and the partition tiles, providing information about boundary contact and exterior interactions. In so doing, the boundary information in particular drastically expands the granularity of what can be discerned from other things.

If anyone knows of relevant literature that suggest the benefits of keeping extra data beyond what is currently used, I would be much obliged.

Similar questions and discussions