8282 tags : pyiceberg-spark:latest
8383 cache-from : type=gha,scope=spark
8484 cache-to : type=gha,mode=max,scope=spark
85+ outputs : type=docker,dest=/tmp/spark-image.tar
8586
8687 - name : Build and cache Hive image
8788 uses : docker/build-push-action@v6
9091 tags : pyiceberg-hive:latest
9192 cache-from : type=gha,scope=hive
9293 cache-to : type=gha,mode=max,scope=hive
94+ outputs : type=docker,dest=/tmp/hive-image.tar
95+
96+ - name : Upload Docker images
97+ uses : actions/upload-artifact@v4
98+ with :
99+ name : docker-images
100+ path : /tmp/*-image.tar
101+ retention-days : 1
102+ compression-level : 0
93103
94104 integration-test :
95105 runs-on : ubuntu-latest
@@ -110,24 +120,16 @@ jobs:
110120 - name : Install
111121 run : make install
112122
113- - name : Set up Docker Buildx
114- uses : docker/setup-buildx-action@v3
115-
116- - name : Build Spark image (from cache)
117- uses : docker/build-push-action@v6
123+ - name : Download Docker images
124+ uses : actions/download-artifact@v4
118125 with :
119- context : dev/spark
120- load : true
121- tags : pyiceberg-spark:latest
122- cache-from : type=gha,scope=spark
126+ name : docker-images
127+ path : /tmp
123128
124- - name : Build Hive image (from cache)
125- uses : docker/build-push-action@v6
126- with :
127- context : dev/hive
128- load : true
129- tags : pyiceberg-hive:latest
130- cache-from : type=gha,scope=hive
129+ - name : Load Docker images
130+ run : |
131+ docker load -i /tmp/spark-image.tar
132+ docker load -i /tmp/hive-image.tar
131133
132134 - name : Run integration tests with coverage
133135 run : COVERAGE=1 make test-integration
0 commit comments