Any information about LakeFS processing with S3 would be useful.
Sharing this problem investigation and discussion for the context: the problem is in the renaming (for S3 filesystem it's copy and delete) after switching to the so called Hadoop ”S3A Committer" - magic committer.
My current opinion that delete may happen before copy due to S3 eventual consistency without special S3Guard configuring - S3Guard setup is not required anymore according to the current state S3 about
strong consistency (e.g.
this article).
Meanwhile the Hadoop
doc describes the next for the magic committer:
However, it has extra requirements of the filesystem
1. The object store must be consistent.
2. The S3A client must be configured to recognize interactions with the magic directories and treat them as a special case.
So, my main questions and any help about LakeFS processing is related to these requirements
1. optionally - is LakeFS object store consistent?
2.
Is there any configuring related to the magic committer and LakeFS as a S3A client?