Paul Singman
spark.conf.set("fs.s3a.bucket.my-repo.access.key", "xxxx")
spark.conf.set("fs.s3a.bucket.my-repo.secret.key", "xxxxxx")
spark.conf.set("fs.s3a.bucket.my-repo.endpoint", "<https://exact-barnacle.lakefs-demo.io>")
spark.conf.set("fs.s3a.path.style.access", "true")
data_path = "/databricks-datasets/amazon/test4K/"
data = spark.read.parquet(data_path)
lakefs_repo = 'my-repo'
lakefs_branch = 'main'
tablename = 'amazon_reviews'
data.write.mode('append').save(f"s3a://{lakefs_repo}/{lakefs_branch}/{tablename}/")
and here’s the error I get:
com.amazonaws.services.securitytoken.model.AWSSecurityTokenServiceException: The security token included in the request is invalid. (Service: AWSSecurityTokenService; Status Code: 403; Error Code: InvalidClientTokenId; Request ID: c9582a88-f367-4a89-bcac-c91579031a14)
is this expected behavior of the playground env or is it user error. Thank you in advance!Yoni Augarten
04/13/2022, 11:41 AMPaul Singman
10.2 (includes Apache Spark 3.2.0, Scala 2.12)
Yoni Augarten
04/13/2022, 12:01 PMYoni Augarten
04/13/2022, 12:13 PMmy-repo
):
spark.hadoop.fs.s3a.bucket.<repo-name>.aws.credentials.provider shaded.databricks.org.apache.hadoop.fs.s3a.TemporaryAWSCredentialsProvider
spark.hadoop.fs.s3a.bucket.<repo-name>.session.token lakefs
Also see the relevant docs.Paul Singman
Paul Singman
com.databricks.s3commit.S3CommitFailedException: java.io.IOException: Bucket my-repo does not exist
Yoni Augarten
04/13/2022, 12:33 PMPaul Singman
Yoni Augarten
04/13/2022, 12:35 PMPaul Singman
Yoni Augarten
04/13/2022, 12:45 PMspark.databricks.delta.multiClusterWrites.enabled false
Paul Singman
Paul Singman
Yoni Augarten
04/13/2022, 1:00 PM