Skip to content

Commit f26cdfa

Browse files
NickLarsenNZadwk67razvan
authored
refactor: Release 24.3 changes (#24)
* feat(stack/hdfs-hbase): upgade zookeeper * refactor(stack/hdfs-hbase): move listenerClass configs to appropriate roles * feat(stack/hdfs-hbase): add listener operator * docs(demo/hbase-hdfs-load-cycling-data): improve readability * docs(demo/hbase-hdfs-load-cycling-data): improve instructions * chore(stack/template): update jupyterlab template * chore(stack/jupyterhub-pyspark-hdfs): update pyspark image * chore(demo/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data): update image tags * chore(demo/hbase-hdfs-load-cycling-data): update image tags * chore(stack/data-lakehouse-iceberg-trino-spark): update productVersion * docs(demo/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data): improve descriptions * docs(demo/hbase-hdfs-load-cycling-data): improve readability * docs(demo/data-lakehouse-iceberg-trino-spark): add install instructions * docs(demo/dairflow-scheduled-job): add install instructions * docs: add install command, move system requirements up, sentence case titles * docs(demo/logging): reorganise docs, update images * docs: add install command, move system requirements up, sentence case titles * chore(demo/logging): bump zookeeper to 3.9.1 * chore(demo/logging): bump testing tools to 0.0.0-dev * chore(stack/jupyterhub-pyspark-hdfs): move hdfs listenerClass, point to new image in notebook, bump zookeeper to 3.9.1 * chore(stack/dual-hive-hdfs-s3): bump trino to 440 * chore(stack/data-lakehouse-iceberg-trino-spark): bump trino to 440 * chore(stack/templates): bump opensearch * chore(demo/nifi-water: bumped products etc. * chore(demo/spark-anomaly-detection): updates to products and dependencies * chore(demo/spark-anomaly-detection): updates to images * all the docs * chore(stack/template): move jupyter template back to python3.9, use notebook on branch * chore(demo/data-lakehouse-iceberg-trino-spark): use 0.0.0-dev images, bump minio image * chore(demo/nifi-water): remove leading space from urls * chore(demo/nifi-earthquake): bump test images * chore(demo/trino-taxi-data): update docs, screenshots, rego rules and stack (trino 442) * chore(demo/trino-taxi-data): restore docs section that still makes sense * chore(demo/signal-processing): bump Nifi, update Grafana and Jupyterlab * bump zookeepers to 3.9.2 * bump airflow to 2.8.1 * chore: update all stacks and images to release 24.3 * fix: image version in jupyter notebook --------- Co-authored-by: Andrew Kenworthy <[email protected]> Co-authored-by: Razvan-Daniel Mihai <[email protected]>
1 parent 9ddda7f commit f26cdfa

File tree

74 files changed

+138
-127
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

74 files changed

+138
-127
lines changed

demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: start-pyspark-job
11-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.3.0
1212
# N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly
1313
# restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps
1414
# below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that

demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: start-date-job
11-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.3.0
1212
# N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly
1313
# restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps
1414
# below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that

demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,11 +9,11 @@ spec:
99
serviceAccountName: demo-serviceaccount
1010
initContainers:
1111
- name: wait-for-kafka
12-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
12+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.3.0
1313
command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/instance=kafka -l app.kubernetes.io/name=kafka"]
1414
containers:
1515
- name: create-nifi-ingestion-job
16-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
16+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1717
command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"]
1818
volumeMounts:
1919
- name: script

demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,11 +12,11 @@ spec:
1212
serviceAccountName: demo-serviceaccount
1313
initContainers:
1414
- name: wait-for-kafka
15-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
15+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.3.0
1616
command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/instance=kafka -l app.kubernetes.io/name=kafka"]
1717
containers:
1818
- name: create-spark-ingestion-job
19-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
19+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.3.0
2020
command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ingestion-job.yaml"]
2121
volumeMounts:
2222
- name: manifest

demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,11 +9,11 @@ spec:
99
serviceAccountName: demo-serviceaccount
1010
initContainers:
1111
- name: wait-for-testdata
12-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
12+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.3.0
1313
command: ["bash", "-c", "echo 'Waiting for job load-test-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-test-data"]
1414
containers:
1515
- name: create-tables-in-trino
16-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
16+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1717
command: ["bash", "-c", "python -u /tmp/script/script.py"]
1818
volumeMounts:
1919
- name: script

demos/data-lakehouse-iceberg-trino-spark/load-test-data.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: load-test-data
11-
image: "bitnami/minio:2022-debian-10"
11+
image: "bitnami/minio:2024-debian-12"
1212
# Please try to order the load jobs from small to large datasets
1313
command: ["bash", "-c", "mc --insecure alias set minio http://minio:9000/ $(cat /minio-s3-credentials/accessKey) $(cat /minio-s3-credentials/secretKey) && cd /tmp && curl -O https://repo.stackable.tech/repository/misc/datasets/open-postcode-geo/open-postcode-geo.csv && mc cp open-postcode-geo.csv minio/staging/house-sales/postcode-geo-lookup/ && rm open-postcode-geo.csv && for year in {2005..2021}; do curl -O https://repo.stackable.tech/repository/misc/datasets/uk-house-sales/uk-house-sales-$year.csv && mc cp uk-house-sales-$year.csv minio/staging/house-sales/house-sales/ && rm uk-house-sales-$year.csv; done && curl -O https://repo.stackable.tech/repository/misc/earthquake-data/earthquakes_1950_to_2022.csv && mc cp earthquakes_1950_to_2022.csv minio/staging/earthquakes/earthquakes/ && rm earthquakes_1950_to_2022.csv && curl -O https://repo.stackable.tech/repository/misc/datasets/e-charging-stations/e-charging-stations-2022-08.csv && mc cp e-charging-stations-2022-08.csv minio/staging/smart-city/e-charging-stations/ && rm e-charging-stations-2022-08.csv && curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/taxi_zone_lookup.csv && mc cp taxi_zone_lookup.csv minio/staging/taxi/taxi-zone-lookup/ && rm taxi_zone_lookup.csv && curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/rate_code_lookup.csv && mc cp rate_code_lookup.csv minio/staging/taxi/rate-code-lookup/ && rm rate_code_lookup.csv && curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/payment_type_lookup.csv && mc cp payment_type_lookup.csv minio/staging/taxi/payment-type-lookup/ && rm payment_type_lookup.csv && for month in 2021-01 2021-02 2021-03 2021-04 2021-05 2021-06 2021-07 2021-08 2021-09 2021-10 2021-11 2021-12 2022-01 2022-02 2022-03 2022-04 2022-05 2022-06; do curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/green_tripdata_$month.parquet && mc cp green_tripdata_$month.parquet minio/staging/taxi/green-tripdata/ && rm green_tripdata_$month.parquet; done && for month in 2015-01 2015-02 2015-03 2015-04 2015-05 2015-06 2015-07 2015-08 2015-09 2015-10 2015-11 2015-12 2016-01 2016-02 2016-03 2016-04 2016-05 2016-06 2016-07 2016-08 2016-09 2016-10 2016-11 2016-12 2017-01 2017-02 2017-03 2017-04 2017-05 2017-06 2017-07 2017-08 2017-09 2017-10 2017-11 2017-12 2018-01 2018-02 2018-03 2018-04 2018-05 2018-06 2018-07 2018-08 2018-09 2018-10 2018-11 2018-12 2019-01 2019-02 2019-03 2019-04 2019-05 2019-06 2019-07 2019-08 2019-09 2019-10 2019-11 2019-12 2020-01 2020-02 2020-03 2020-04 2020-05 2020-06 2020-07 2020-08 2020-09 2020-10 2020-11 2020-12 2021-01 2021-02 2021-03 2021-04 2021-05 2021-06 2021-07 2021-08 2021-09 2021-10 2021-11 2021-12 2022-01 2022-02 2022-03 2022-04 2022-05 2022-06 2022-07 2022-08 2022-09 2022-10 2022-11 2022-12 2023-01 2023-02 2023-03 2023-04; do curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/yellow_tripdata_$month.parquet && mc cp yellow_tripdata_$month.parquet minio/staging/taxi/yellow-tripdata/ && rm yellow_tripdata_$month.parquet; done && for month in 2020-09 2020-10 2020-11 2020-12 2021-01 2021-02 2021-03 2021-04 2021-05 2021-06 2021-07 2021-08 2021-09 2021-10 2021-11 2021-12 2022-01 2022-02 2022-03 2022-04 2022-05 2022-06; do curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/fhvhv_tripdata_$month.parquet && mc cp fhvhv_tripdata_$month.parquet minio/staging/taxi/fhvhv-tripdata/ && rm fhvhv_tripdata_$month.parquet; done"]
1414
volumeMounts:

demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: setup-superset
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"]
1313
volumeMounts:
1414
- name: script

demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ spec:
99
spec:
1010
containers:
1111
- name: create-hfile-and-import-to-hbase
12-
image: docker.stackable.tech/stackable/hbase:2.4.17-stackable23.11.0
12+
image: docker.stackable.tech/stackable/hbase:2.4.17-stackable24.3.0
1313
env:
1414
- name: HADOOP_USER_NAME
1515
value: stackable

demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: distcp-cycling-data
11-
image: docker.stackable.tech/stackable/hadoop:3.3.6-stackable23.11.0
11+
image: docker.stackable.tech/stackable/hadoop:3.3.6-stackable24.3.0
1212
env:
1313
- name: HADOOP_USER_NAME
1414
value: stackable

demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/Dockerfile

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
1-
# docker build -t docker.stackable.tech/demos/pyspark-k8s-with-scikit-learn:3.3.0-stackable23.4 .
1+
# docker build -t docker.stackable.tech/demos/pyspark-k8s-with-scikit-learn:3.3.0-stackable0.0.0-dev .
22

3-
FROM docker.stackable.tech/stackable/pyspark-k8s:3.3.0-stackable23.4
3+
FROM docker.stackable.tech/stackable/pyspark-k8s:3.5.0-stackable24.3.0
44

55
COPY requirements.txt .
66

demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: load-ny-taxi-data
11-
image: docker.stackable.tech/stackable/hadoop:3.3.6-stackable23.11.0
11+
image: docker.stackable.tech/stackable/hadoop:3.3.6-stackable24.3.0
1212
command: ["bash", "-c", "/stackable/hadoop/bin/hdfs dfs -mkdir -p /ny-taxi-data/raw \
1313
&& cd /tmp \
1414
&& for month in 2020-09; do \

demos/logging/zookeeper.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ metadata:
55
name: simple-zk
66
spec:
77
image:
8-
productVersion: 3.8.3
8+
productVersion: 3.9.2
99
clusterConfig:
1010
vectorAggregatorConfigMapName: vector-aggregator-discovery
1111
servers:

demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: create-druid-ingestion-job
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor"]
1313
volumeMounts:
1414
- name: ingestion-job-spec

demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: create-nifi-ingestion-job
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"]
1313
volumeMounts:
1414
- name: script

demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: setup-superset
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"]
1313
volumeMounts:
1414
- name: script

demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2075,7 +2075,7 @@
20752075
</entry>
20762076
<entry>
20772077
<key>Remote URL</key>
2078-
<value> https://www.pegelonline.wsv.de/webservices/rest-api/v2/stations.json</value>
2078+
<value>https://www.pegelonline.wsv.de/webservices/rest-api/v2/stations.json</value>
20792079
</entry>
20802080
<entry>
20812081
<key>disable-http2</key>
@@ -3370,7 +3370,7 @@
33703370
</entry>
33713371
<entry>
33723372
<key>Remote URL</key>
3373-
<value> https://www.pegelonline.wsv.de/webservices/rest-api/v2/stations.json</value>
3373+
<value>https://www.pegelonline.wsv.de/webservices/rest-api/v2/stations.json</value>
33743374
</entry>
33753375
<entry>
33763376
<key>disable-http2</key>
@@ -3855,7 +3855,7 @@
38553855
</entry>
38563856
<entry>
38573857
<key>Remote URL</key>
3858-
<value> https://www.pegelonline.wsv.de/webservices/rest-api/v2/stations.json</value>
3858+
<value>https://www.pegelonline.wsv.de/webservices/rest-api/v2/stations.json</value>
38593859
</entry>
38603860
<entry>
38613861
<key>disable-http2</key>

demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: create-druid-ingestion-job
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/stations-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-compaction-job-spec.json https://druid-coordinator:8281/druid/coordinator/v1/config/compaction"]
1313
volumeMounts:
1414
- name: ingestion-job-spec

demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: create-nifi-ingestion-job
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"]
1313
volumeMounts:
1414
- name: script

demos/nifi-kafka-druid-water-level-data/setup-superset.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: setup-superset
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"]
1313
volumeMounts:
1414
- name: script
Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
1-
# docker build -f ./Dockerfile-nifi -t docker.stackable.tech/demos/nifi:1.21.0-postgresql .
1+
# docker build -f ./Dockerfile-nifi -t docker.stackable.tech/demos/nifi:1.25.0-postgresql .
22

3-
FROM docker.stackable.tech/stackable/nifi:1.21.0-stackable23.7
3+
FROM docker.stackable.tech/stackable/nifi:1.25.0-stackable24.3.0
44

55
RUN curl --fail -o /stackable/nifi/postgresql-42.6.0.jar "https://repo.stackable.tech/repository/misc/postgresql-timescaledb/postgresql-42.6.0.jar"

demos/signal-processing/create-nifi-ingestion-job.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,13 +9,13 @@ spec:
99
serviceAccountName: demo-serviceaccount
1010
initContainers:
1111
- name: wait-for-timescale-job
12-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
12+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.3.0
1313
command: ["bash", "-c", "echo 'Waiting for timescaleDB tables to be ready'
1414
&& kubectl wait --for=condition=complete job/create-timescale-tables-job"
1515
]
1616
containers:
1717
- name: create-nifi-ingestion-job
18-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
18+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1919
command: ["bash", "-c", "export PGPASSWORD=$(cat /timescale-admin-credentials/password) && \
2020
curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/DownloadAndWriteToDB.xml && \
2121
sed -i \"s/PLACEHOLDERPGPASSWORD/$PGPASSWORD/g\" DownloadAndWriteToDB.xml && \

demos/signal-processing/create-timescale-tables.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ spec:
99
serviceAccountName: demo-serviceaccount
1010
initContainers:
1111
- name: wait-for-timescale
12-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
12+
image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.3.0
1313
command: ["bash", "-c", "echo 'Waiting for timescaleDB to be ready'
1414
&& kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=postgresql-timescaledb"
1515
]

demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml

Lines changed: 10 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -8,11 +8,11 @@ spec:
88
spec:
99
initContainers:
1010
- name: wait-for-testdata
11-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "echo 'Waiting for job load-ny-taxi-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-ny-taxi-data"]
1313
containers:
1414
- name: create-spark-anomaly-detection-job
15-
image: docker.stackable.tech/stackable/tools:1.0.0-stackable23.11.0
15+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1616
command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ad-job.yaml"]
1717
volumeMounts:
1818
- name: manifest
@@ -36,15 +36,16 @@ data:
3636
metadata:
3737
name: spark-ad
3838
spec:
39-
version: "1.0"
4039
sparkImage:
41-
custom: docker.stackable.tech/demos/pyspark-k8s-with-kafka-and-iceberg:3.3.0-stackable0.2.0
42-
productVersion: 3.3.0
40+
productVersion: 3.5.0
4341
mode: cluster
4442
mainApplicationFile: local:///spark-scripts/spark-ad.py
4543
deps:
44+
packages:
45+
- org.apache.iceberg:iceberg-spark-runtime-3.5_2.12:1.5.0
46+
- org.apache.spark:spark-sql-kafka-0-10_2.12:3.5.0
4647
requirements:
47-
- scikit-learn==0.24.2
48+
- scikit-learn==1.4.0
4849
s3connection:
4950
reference: minio
5051
volumes:
@@ -90,6 +91,9 @@ data:
9091
spark.sql.catalog.prediction: org.apache.iceberg.spark.SparkCatalog
9192
spark.sql.catalog.prediction.type: hive
9293
spark.sql.catalog.prediction.uri: thrift://hive-iceberg:9083
94+
# AFAIK This can be removed in 24.3
95+
#spark.driver.userClassPathFirst: "false"
96+
#spark.executor.userClassPathFirst: "false"
9397
---
9498
apiVersion: v1
9599
kind: ConfigMap

demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: setup-superset
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"]
1313
volumeMounts:
1414
- name: script

demos/trino-taxi-data/create-table-in-trino.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ spec:
88
spec:
99
containers:
1010
- name: create-ny-taxi-data-table-in-trino
11-
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable23.11.0
11+
image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.3.0
1212
command: ["bash", "-c", "python -u /tmp/script/script.py"]
1313
volumeMounts:
1414
- name: script

0 commit comments

Comments
 (0)