Skip to content

Commit

Permalink
updating docs, fixing jar version
Browse files Browse the repository at this point in the history
  • Loading branch information
cugni committed Dec 7, 2021
1 parent e01f5cb commit d9bd04a
Show file tree
Hide file tree
Showing 5 changed files with 10 additions and 10 deletions.
2 changes: 1 addition & 1 deletion CONTRIBUTING.md
Original file line number Diff line number Diff line change
Expand Up @@ -88,7 +88,7 @@ For example:
sbt assembly

$SPARK_HOME/bin/spark-shell \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.1.0.jar \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.2.0.jar \
--conf spark.sql.extensions=io.qbeast.spark.internal.QbeastSparkSessionExtension \
--packages io.delta:delta-core_2.12:1.0.0
```
Expand Down
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -82,7 +82,7 @@ You can find it in the [Packages section](https://github.com/orgs/Qbeast-io/pack

```bash
$SPARK_HOME/bin/spark-shell \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.1.0.jar \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.2.0.jar \
--conf spark.sql.extensions=io.qbeast.spark.internal.QbeastSparkSessionExtension \
--packages io.delta:delta-core_2.12:1.0.0
```
Expand Down
6 changes: 3 additions & 3 deletions docs/CloudStorages.md
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ Amazon Web Services S3 does not work with Hadoop 2.7. For this provider you'll n
- If you are using a **public** bucket:
```bash
$SPARK_HOME/bin/spark-shell \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.1.0.jar \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.2.0.jar \
--conf spark.sql.extensions=io.qbeast.spark.internal.QbeastSparkSessionExtension \
--conf spark.hadoop.fs.s3a.aws.credentials.provider=org.apache.hadoop.fs.s3a.AnonymousAWSCredentialsProvider \
--packages io.delta:delta-core_2.12:1.0.0,\
Expand All @@ -42,7 +42,7 @@ $SPARK_HOME/bin/spark-shell \
- If you are using **private** buckets:
```bash
$SPARK_HOME/bin/spark-shell \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.1.0.jar \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.2.0.jar \
--conf spark.sql.extensions=io.qbeast.spark.internal.QbeastSparkSessionExtension \
--conf spark.hadoop.fs.s3a.access.key=${AWS_ACCESS_KEY_ID} \
--conf spark.hadoop.fs.s3a.secret.key=${AWS_SECRET_ACCESS_KEY} \
Expand All @@ -60,7 +60,7 @@ recommend using the latest. Remember that vanilla parquet format may not work in
- An example config setup follows:
```bash
$SPARK_HOME/bin/spark-shell \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.1.0.jar \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.2.0.jar \
--conf spark.hadoop.fs.azure.account.key.blobqsql.blob.core.windows.net="${AZURE_BLOB_STORAGE_KEY}" \
--conf spark.hadoop.fs.AbstractFileSystem.wasb.impl=org.apache.hadoop.fs.azure.Wasb \
--conf spark.sql.extensions=io.qbeast.spark.internal.QbeastSparkSessionExtension \
Expand Down
6 changes: 3 additions & 3 deletions docs/Quickstart.md
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ Inside the project folder, launch a spark-shell with the required **dependencies
```bash
$SPARK_HOME/bin/spark-shell \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.1.0.jar \
--jars ./target/scala-2.12/qbeast-spark-assembly-0.2.0.jar \
--conf spark.sql.extensions=io.qbeast.spark.internal.QbeastSparkSessionExtension \
--conf spark.hadoop.fs.s3a.aws.credentials.provider=org.apache.hadoop.fs.s3a.AnonymousAWSCredentialsProvider \
--packages io.delta:delta-core_2.12:1.0.0,\
Expand Down Expand Up @@ -48,12 +48,12 @@ Indexing the data with the desired columns, in this case `ss_cdemo_sk` and `ss_c
```scala
val qbeastTablePath = "/tmp/qbeast-test-data/qtable"

parquetDf.write
(parquetDf.write
.mode("overwrite")
.format("qbeast") // Saving the dataframe in a qbeast datasource
.option("columnsToIndex", "ss_cdemo_sk,ss_cdemo_sk") // Indexing the table
.option("cubeSize", 300000) // The desired number of records of the resulting files/cubes. Default is 100000
.save(qbeastTablePath)
.save(qbeastTablePath))
```

## Sampling
Expand Down
4 changes: 2 additions & 2 deletions docs/sample_pushdown_demo.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -129,7 +129,7 @@
"\n",
"spark = (SparkSession.builder\n",
" .master(\"local[*]\")\n",
" .config(\"spark.jars\", \"../target/scala-2.12/qbeast-spark-assembly-0.1.0.jar\")\n",
" .config(\"spark.jars\", \"../target/scala-2.12/qbeast-spark-assembly-0.2.0.jar\")\n",
" .config(\"spark.sql.extensions\", \"io.qbeast.spark.internal.QbeastSparkSessionExtension\")\n",
" .config(\"fs.s3a.aws.credentials.provider\", \"org.apache.hadoop.fs.s3a.AnonymousAWSCredentialsProvider\")\n",
" .config(\"spark.jars.packages\", deps)\n",
Expand Down Expand Up @@ -469,4 +469,4 @@
},
"nbformat": 4,
"nbformat_minor": 5
}
}

0 comments on commit d9bd04a

Please sign in to comment.