Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 6 additions & 2 deletions docker/thirdparties/docker-compose/iceberg/iceberg.yaml.tpl
Original file line number Diff line number Diff line change
Expand Up @@ -30,15 +30,19 @@ services:
- ./data/output/spark-warehouse:/home/iceberg/warehouse
- ./data/output/spark-notebooks:/home/iceberg/notebooks/notebooks
- ./data:/mnt/data
- ./spark-init.sql:/mnt/spark-init.sql
- ./spark-init-iceberg.sql:/mnt/spark-init-iceberg.sql
- ./spark-init-paimon.sql:/mnt/spark-init-paimon.sql
- ./spark-defaults.conf:/opt/spark/conf/spark-defaults.conf
- ./data/input/jars/paimon-spark-3.5-0.8.0.jar:/opt/spark/jars/paimon-spark-3.5-0.8.0.jar
- ./data/input/jars/paimon-s3-0.8.0.jar:/opt/spark/jars/paimon-s3-0.8.0.jar
environment:
- AWS_ACCESS_KEY_ID=admin
- AWS_SECRET_ACCESS_KEY=password
- AWS_REGION=us-east-1
entrypoint: >
/bin/sh -c "
spark-sql -f /mnt/spark-init.sql 2>&1;
spark-sql --conf spark.sql.extensions=org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions -f /mnt/spark-init-iceberg.sql 2>&1;
spark-sql --conf spark.sql.extensions=org.apache.paimon.spark.extensions.PaimonSparkSessionExtensions -f /mnt/spark-init-paimon.sql 2>&1;
tail -f /dev/null
"
networks:
Expand Down
11 changes: 9 additions & 2 deletions docker/thirdparties/docker-compose/iceberg/spark-defaults.conf
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,6 @@

# Example:
spark.sql.session.timeZone Asia/Shanghai
spark.sql.extensions org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions
spark.sql.catalog.demo org.apache.iceberg.spark.SparkCatalog
spark.sql.catalog.demo.type rest
spark.sql.catalog.demo.uri http://rest:8181
Expand All @@ -31,4 +30,12 @@ spark.sql.defaultCatalog demo
spark.eventLog.enabled true
spark.eventLog.dir /home/iceberg/spark-events
spark.history.fs.logDirectory /home/iceberg/spark-events
spark.sql.catalogImplementation in-memory
spark.sql.catalogImplementation in-memory

# paimon
spark.sql.catalog.paimon org.apache.paimon.spark.SparkCatalog
spark.sql.catalog.paimon.warehouse s3://warehouse/wh
spark.sql.catalog.paimon.s3.endpoint http://minio:9000
spark.sql.catalog.paimon.s3.access-key admin
spark.sql.catalog.paimon.s3.secret-key password
spark.sql.catalog.paimon.s3.region us-east-1
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
-- create database if not exists paimon.test_paimon_db;
19 changes: 14 additions & 5 deletions docker/thirdparties/run-thirdparties-docker.sh
Original file line number Diff line number Diff line change
Expand Up @@ -393,17 +393,26 @@ fi

if [[ "${RUN_ICEBERG}" -eq 1 ]]; then
# iceberg
ICEBERG_DIR=${ROOT}/docker-compose/iceberg
cp "${ROOT}"/docker-compose/iceberg/iceberg.yaml.tpl "${ROOT}"/docker-compose/iceberg/iceberg.yaml
cp "${ROOT}"/docker-compose/iceberg/entrypoint.sh.tpl "${ROOT}"/docker-compose/iceberg/entrypoint.sh
sed -i "s/doris--/${CONTAINER_UID}/g" "${ROOT}"/docker-compose/iceberg/iceberg.yaml
sed -i "s/doris--/${CONTAINER_UID}/g" "${ROOT}"/docker-compose/iceberg/entrypoint.sh
sudo docker compose -f "${ROOT}"/docker-compose/iceberg/iceberg.yaml --env-file "${ROOT}"/docker-compose/iceberg/iceberg.env down
sudo rm -rf "${ROOT}"/docker-compose/iceberg/data
if [[ "${STOP}" -ne 1 ]]; then
wget -P "${ROOT}"/docker-compose/iceberg https://"${s3BucketName}.${s3Endpoint}"/regression/datalake/pipeline_data/iceberg_data.zip
sudo unzip -d "${ROOT}"/docker-compose/iceberg -q ${ROOT}/docker-compose/iceberg/iceberg_data.zip
sudo mv "${ROOT}"/docker-compose/iceberg/iceberg_data "${ROOT}"/docker-compose/iceberg/data
sudo rm -rf ${ROOT}/docker-compose/iceberg/iceberg_data.zip
if [[ ! -d "${ICEBERG_DIR}/data" ]]; then
echo "${ICEBERG_DIR}/data does not exist"
cd "${ICEBERG_DIR}" \
&& rm -f iceberg_data.zip \
&& wget -P "${ROOT}"/docker-compose/iceberg https://"${s3BucketName}.${s3Endpoint}"/regression/datalake/pipeline_data/iceberg_data.zip \
&& sudo unzip iceberg_data.zip \
&& sudo mv iceberg_data data \
&& sudo rm -rf iceberg_data.zip
cd -
else
echo "${ICEBERG_DIR}/data exist, continue !"
fi

sudo docker compose -f "${ROOT}"/docker-compose/iceberg/iceberg.yaml --env-file "${ROOT}"/docker-compose/iceberg/iceberg.env up -d
fi
fi
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -121,6 +121,11 @@ private Type paimonPrimitiveTypeToDorisType(org.apache.paimon.types.DataType dat
if (scale > 6) {
scale = 6;
}
} else if (dataType instanceof org.apache.paimon.types.LocalZonedTimestampType) {
scale = ((org.apache.paimon.types.LocalZonedTimestampType) dataType).getPrecision();
if (scale > 6) {
scale = 6;
}
}
return ScalarType.createDatetimeV2Type(scale);
case ARRAY:
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,29 @@
-- This file is automatically generated. You should know what you did if you want to edit this
-- !q01 --
2023-03-08
2023-03-09

-- !q02 --
1996-05-06
1996-05-06
1997-05-18
1997-05-18

-- !q03 --
1996-05-06 MAIL
1997-05-18 MAIL

-- !q04 --
1996-05-01 MAI
1996-05-06 MAI
1996-05-06 MAIL
1997-05-18 MAI
1997-05-18 MAIL
1997-05-19 MAI

-- !q05 --
2023-03-07T20:35:59.123456
2023-03-07T20:35:59.123456
2023-03-07T20:36
2023-03-07T20:37:59

This file was deleted.

This file was deleted.

Loading