|
9 | 9 | "name": "stderr", |
10 | 10 | "output_type": "stream", |
11 | 11 | "text": [ |
12 | | - "25/03/22 00:13:18 WARN SparkSession: Using an existing Spark session; only runtime SQL configurations will take effect.\n" |
| 12 | + "25/03/23 21:11:07 WARN SparkSession: Using an existing Spark session; only runtime SQL configurations will take effect.\n" |
13 | 13 | ] |
14 | 14 | } |
15 | 15 | ], |
|
28 | 28 | "data": { |
29 | 29 | "text/plain": [ |
30 | 30 | "[('spark.sql.catalog.nessie', 'org.apache.iceberg.spark.SparkCatalog'),\n", |
31 | | - " ('spark.app.submitTime', '1742601784570'),\n", |
32 | 31 | " ('spark.sql.catalog.nessie.uri', 'http://nessie:19120/api/v1'),\n", |
33 | 32 | " ('spark.hadoop.fs.s3a.path.style.access', 'true'),\n", |
34 | 33 | " ('spark.sql.catalog.nessie.ref', 'main'),\n", |
35 | | - " ('spark.app.startTime', '1742601784818'),\n", |
36 | | - " ('spark.driver.port', '39441'),\n", |
37 | 34 | " ('spark.sql.warehouse.dir',\n", |
38 | 35 | " 'file:/home/iceberg/notebooks/notebooks/spark-warehouse'),\n", |
| 36 | + " ('spark.app.submitTime', '1742764264501'),\n", |
39 | 37 | " ('spark.hadoop.fs.s3a.access.key', 'minioadmin'),\n", |
40 | 38 | " ('spark.sql.catalog.nessie.s3.path-style-access', 'true'),\n", |
41 | 39 | " ('spark.serializer.objectStreamReset', '100'),\n", |
|
50 | 48 | " 'file:///root/.ivy2/jars/org.apache.iceberg_iceberg-spark-runtime-3.5_2.12-1.8.1.jar,file:///root/.ivy2/jars/org.projectnessie.nessie-integrations_nessie-spark-extensions-3.5_2.12-0.103.2.jar'),\n", |
51 | 49 | " ('spark.driver.extraJavaOptions',\n", |
52 | 50 | " '-Djava.net.preferIPv6Addresses=false -XX:+IgnoreUnrecognizedVMOptions --add-opens=java.base/java.lang=ALL-UNNAMED --add-opens=java.base/java.lang.invoke=ALL-UNNAMED --add-opens=java.base/java.lang.reflect=ALL-UNNAMED --add-opens=java.base/java.io=ALL-UNNAMED --add-opens=java.base/java.net=ALL-UNNAMED --add-opens=java.base/java.nio=ALL-UNNAMED --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.util.concurrent.atomic=ALL-UNNAMED --add-opens=java.base/jdk.internal.ref=ALL-UNNAMED --add-opens=java.base/sun.nio.ch=ALL-UNNAMED --add-opens=java.base/sun.nio.cs=ALL-UNNAMED --add-opens=java.base/sun.security.action=ALL-UNNAMED --add-opens=java.base/sun.util.calendar=ALL-UNNAMED --add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED -Djdk.reflect.useDirectMethodHandle=false'),\n", |
53 | | - " ('spark.app.id', 'local-1742601785395'),\n", |
| 51 | + " ('spark.app.startTime', '1742764264699'),\n", |
54 | 52 | " ('spark.sql.catalog.nessie.s3.endpoint', 'http://minio:9000'),\n", |
| 53 | + " ('spark.app.id', 'local-1742764265168'),\n", |
55 | 54 | " ('spark.repl.local.jars',\n", |
56 | 55 | " 'file:///root/.ivy2/jars/org.apache.iceberg_iceberg-spark-runtime-3.5_2.12-1.8.1.jar,file:///root/.ivy2/jars/org.projectnessie.nessie-integrations_nessie-spark-extensions-3.5_2.12-0.103.2.jar'),\n", |
57 | 56 | " ('spark.executor.id', 'driver'),\n", |
58 | | - " ('spark.driver.host', '876c347e2cee'),\n", |
59 | 57 | " ('spark.app.name', 'PySparkShell'),\n", |
60 | 58 | " ('spark.hadoop.fs.s3a.impl', 'org.apache.hadoop.fs.s3a.S3AFileSystem'),\n", |
| 59 | + " ('spark.driver.host', '137a0ca84ad6'),\n", |
61 | 60 | " ('spark.sql.catalogImplementation', 'hive'),\n", |
62 | 61 | " ('spark.sql.catalog.nessie.io-impl', 'org.apache.iceberg.aws.s3.S3FileIO'),\n", |
63 | 62 | " ('spark.submit.pyFiles',\n", |
|
66 | 65 | " ('spark.executor.extraJavaOptions',\n", |
67 | 66 | " '-Djava.net.preferIPv6Addresses=false -XX:+IgnoreUnrecognizedVMOptions --add-opens=java.base/java.lang=ALL-UNNAMED --add-opens=java.base/java.lang.invoke=ALL-UNNAMED --add-opens=java.base/java.lang.reflect=ALL-UNNAMED --add-opens=java.base/java.io=ALL-UNNAMED --add-opens=java.base/java.net=ALL-UNNAMED --add-opens=java.base/java.nio=ALL-UNNAMED --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.util.concurrent.atomic=ALL-UNNAMED --add-opens=java.base/jdk.internal.ref=ALL-UNNAMED --add-opens=java.base/sun.nio.ch=ALL-UNNAMED --add-opens=java.base/sun.nio.cs=ALL-UNNAMED --add-opens=java.base/sun.security.action=ALL-UNNAMED --add-opens=java.base/sun.util.calendar=ALL-UNNAMED --add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED -Djdk.reflect.useDirectMethodHandle=false'),\n", |
68 | 67 | " ('spark.sql.catalog.nessie.warehouse', 's3a://warehouse'),\n", |
| 68 | + " ('spark.app.initial.jar.urls',\n", |
| 69 | + " 'spark://137a0ca84ad6:46315/jars/org.projectnessie.nessie-integrations_nessie-spark-extensions-3.5_2.12-0.103.2.jar,spark://137a0ca84ad6:46315/jars/org.apache.iceberg_iceberg-spark-runtime-3.5_2.12-1.8.1.jar'),\n", |
69 | 70 | " ('spark.jars',\n", |
70 | 71 | " 'file:///root/.ivy2/jars/org.apache.iceberg_iceberg-spark-runtime-3.5_2.12-1.8.1.jar,file:///root/.ivy2/jars/org.projectnessie.nessie-integrations_nessie-spark-extensions-3.5_2.12-0.103.2.jar'),\n", |
71 | 72 | " ('spark.hadoop.fs.s3a.endpoint', 'http://minio:9000'),\n", |
| 73 | + " ('spark.driver.port', '46315'),\n", |
| 74 | + " ('spark.sql.catalog.nessie.type', 'nessie'),\n", |
72 | 75 | " ('spark.files',\n", |
73 | 76 | " 'file:///root/.ivy2/jars/org.apache.iceberg_iceberg-spark-runtime-3.5_2.12-1.8.1.jar,file:///root/.ivy2/jars/org.projectnessie.nessie-integrations_nessie-spark-extensions-3.5_2.12-0.103.2.jar'),\n", |
74 | | - " ('spark.sql.catalog.nessie.type', 'nessie'),\n", |
75 | | - " ('spark.app.initial.jar.urls',\n", |
76 | | - " 'spark://876c347e2cee:39441/jars/org.apache.iceberg_iceberg-spark-runtime-3.5_2.12-1.8.1.jar,spark://876c347e2cee:39441/jars/org.projectnessie.nessie-integrations_nessie-spark-extensions-3.5_2.12-0.103.2.jar'),\n", |
77 | 77 | " ('spark.ui.showConsoleProgress', 'true')]" |
78 | 78 | ] |
79 | 79 | }, |
|
95 | 95 | "name": "stderr", |
96 | 96 | "output_type": "stream", |
97 | 97 | "text": [ |
98 | | - "25/03/22 00:13:22 WARN HiveConf: HiveConf of name hive.stats.jdbc.timeout does not exist\n", |
99 | | - "25/03/22 00:13:22 WARN HiveConf: HiveConf of name hive.stats.retries.wait does not exist\n", |
100 | | - "25/03/22 00:13:22 WARN ObjectStore: Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 2.3.0\n", |
101 | | - "25/03/22 00:13:22 WARN ObjectStore: setMetaStoreSchemaVersion called but recording version is disabled: version = 2.3.0, comment = Set by MetaStore [email protected]\n", |
102 | | - "25/03/22 00:13:23 WARN ObjectStore: Failed to get database global_temp, returning NoSuchObjectException\n" |
| 98 | + "25/03/23 21:11:15 WARN HiveConf: HiveConf of name hive.stats.jdbc.timeout does not exist\n", |
| 99 | + "25/03/23 21:11:15 WARN HiveConf: HiveConf of name hive.stats.retries.wait does not exist\n", |
| 100 | + "25/03/23 21:11:16 WARN ObjectStore: Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 2.3.0\n", |
| 101 | + "25/03/23 21:11:16 WARN ObjectStore: setMetaStoreSchemaVersion called but recording version is disabled: version = 2.3.0, comment = Set by MetaStore [email protected]\n", |
| 102 | + "25/03/23 21:11:16 WARN ObjectStore: Failed to get database global_temp, returning NoSuchObjectException\n" |
103 | 103 | ] |
104 | 104 | }, |
105 | 105 | { |
|
138 | 138 | } |
139 | 139 | ], |
140 | 140 | "source": [ |
141 | | - "spark.sql(\"create namespace default\")" |
| 141 | + "spark.sql(\"CREATE BRANCH te\")" |
142 | 142 | ] |
143 | 143 | }, |
144 | 144 | { |
|
0 commit comments