Robin Moffatt
03/23/2023, 4:12 PMAriel Shaqed (Scolnicov)
03/24/2023, 1:12 PMWe have no reason to believe that the exposed key was abused and took this action out of an abundance of caution.
Also includes instructions how to tell your SSH client to drop the compromised old public key so you can verify and start trusting the new key. AFAIK your SSH client should not normally be using this key, you are most likely using an EC* key instead.
Niro
03/28/2023, 12:30 PMAriel Shaqed (Scolnicov)
04/03/2023, 7:57 AM.../_lakefs/logs/gc/expired_objects
.
THANKS!Ariel Shaqed (Scolnicov)
04/03/2023, 10:52 AM_"With great scale comes big data_"And we're getting there. I think it's time for pack files like Git uses, so I made up this issue to design how we can do them. Packs serve as snapshots in big data (or cubes if anyone still remembers OLAP): in this case, I would like us to hold many commits together in a single "file". The issue has a sketch, and I think we need to think about this direction. I believe it can help us with: • GC on large repositories with many commits • lakectl blame and other log-by-path operations • Even FindMergeBase when the base is distant. It won't be a short sprint, but I'd like us to think about this as a possible next step for lakeFS scalability.
Robin Moffatt
04/05/2023, 6:32 AM<https://github.com/treeverse/lakeFS/blob/master/docker-compose.yaml>
is for these days? It looks like it was added a while ago.
I'd like to [re]move it so that new users aren't confused between it and the quickstart Docker Compose, but wanted to check first before doing so.Guy Hardonag
04/05/2023, 6:42 AMOz Katz
04/05/2023, 9:49 AMGuy Hardonag
04/19/2023, 8:29 AMGuy Hardonag
04/19/2023, 10:06 AMBarak Amar
04/24/2023, 8:24 AMRobin Moffatt
04/25/2023, 10:02 AMAKIAIOSFODNN7EXAMPLE
and wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY
as example access&secret keys respectively, throughout the documentation, Everything Bagel, etc.
If we replaced these with AKIA-EXAMPLE-KEY
and EXAMPLE-SECRET
it would make documentation more accessible by being more readable (the brain doesn't have to grok a bunch of random chars to realise it's just a dummy placeholder). More accessible == easier for users to get to grips with lakeFS and the samples and documentation.
Before I do a PR with a global search & replace for formal review, I wanted to check if there were any immediate objections or gotchas to consider ?
thanks 🙂Yoni Augarten
04/27/2023, 8:46 AMAriel Shaqed (Scolnicov)
04/27/2023, 4:50 PMAriel Shaqed (Scolnicov)
04/30/2023, 7:51 AMOz Katz
04/30/2023, 9:50 AMManoj Babu
04/30/2023, 8:22 PMAriel Shaqed (Scolnicov)
05/02/2023, 3:56 PMAriel Shaqed (Scolnicov)
05/02/2023, 3:57 PMRobin Moffatt
05/05/2023, 7:10 AMDockerfile
and not being totally familiar with them beyond the basics wondered if the duplication of some of these lines is deliberate and if so what purpose it serves? Thanks.
```EXPOSE 8000/tcp
# Setup user
RUN addgroup -S lakefs && adduser -S lakefs -G lakefs
USER lakefs
WORKDIR /home/lakefs
ENTRYPOINT ["/app/lakefs"]
CMD ["run"]```I'm planning a PR on the file so will submit a second one to tidy things up, if necessary
Manoj Babu
05/07/2023, 7:23 AMVaibhav Kumar
05/07/2023, 6:22 PMManoj Babu
05/08/2023, 9:20 AMhttps://github.com/treeverse/lakeFS/blob/5d237c497f9c2a51d665d3a96075f7c783f7c928/docs/assets/logo.png▾
giphy
05/10/2023, 10:20 AMAriel Shaqed (Scolnicov)
05/10/2023, 10:21 AMBarak Amar
05/18/2023, 9:34 AMVaibhav Kumar
05/19/2023, 7:06 PMspark-shell --conf spark.hadoop.fs.s3a.access.key=minioadmin\
--conf spark.hadoop.fs.s3a.secret.key=minioadmin\
--conf spark.hadoop.fs.s3a.endpoint=<http://127.0.0.1:9090>\
--conf spark.hadoop.fs.lakefs.impl=io.lakefs.LakeFSFileSystem\
--conf spark.hadoop.fs.lakefs.access.key=AKIAIOSFODNN7EXAMPLE\
--conf spark.hadoop.fs.lakefs.secret.key='wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY'\
--conf spark.hadoop.fs.lakefs.endpoint=<http://localhost:8000/api/v1>\
—jars /Users/simar/lakeFS/clients/hadoopfs/target/hadoop-lakefs-0.1.0.jar\
io.lakefs.LakeFSFileSystem
While reading it I am getting the below error
scala> val df = spark.read.parquet("<lakefs://example/main/sample1.json>")
23/05/20 00:24:42 WARN FileStreamSink: Assume no metadata directory. Error while looking for metadata directory in the path: <lakefs://example/main/sample1.json>.
java.lang.RuntimeException: java.lang.ClassNotFoundException: Class io.lakefs.LakeFSFileSystem not found
Does anyone know what could be causing this?Ariel Shaqed (Scolnicov)
05/26/2023, 4:23 AMRobin Moffatt
05/30/2023, 4:48 PM-duckdb
Docker image and reference in docs quickstart each time.
I'd like to help towards this where possible, but not sure where to start 🙂
(at the moment there is no treeverse/lakefs:0.101.0-duckdb
, only for 0.100)Niro
05/31/2023, 8:46 AM