@@ -127,19 +127,7 @@ def build_bundles(event, context):
127
127
128
128
base_url = client .server_info ()["capabilities" ]["attachments" ]["base_url" ]
129
129
130
- existing_bundle_timestamp = get_modified_timestamp (
131
- f"{ base_url } { DESTINATION_FOLDER } /changesets.zip"
132
- )
133
- if existing_bundle_timestamp is None :
134
- print ("No previous bundle found" ) # Should only happen once.
135
- existing_bundle_timestamp = - 1
136
-
137
130
all_changesets = fetch_all_changesets (client )
138
- highest_timestamp = max (c ["timestamp" ] for c in all_changesets )
139
-
140
- if existing_bundle_timestamp >= highest_timestamp :
141
- print ("Existing bundle up-to-date. Nothing to do." )
142
- return
143
131
144
132
# Build all archives in temp directory.
145
133
tmp_dir = tempfile .mkdtemp ()
@@ -148,26 +136,6 @@ def build_bundles(event, context):
148
136
bundles_to_upload = []
149
137
bundles_to_delete = []
150
138
151
- write_zip (
152
- "changesets.zip" ,
153
- [
154
- ("{metadata[bucket]}--{metadata[id]}.json" .format (** changeset ), json .dumps (changeset ))
155
- for changeset in all_changesets
156
- ],
157
- )
158
- bundles_to_upload .append ("changesets.zip" )
159
-
160
- # Build a bundle for collections that are marked with "startup" flag.
161
- write_zip (
162
- "startup.zip" ,
163
- [
164
- ("{metadata[bucket]}--{metadata[id]}.json" .format (** changeset ), json .dumps (changeset ))
165
- for changeset in all_changesets
166
- if "startup" in changeset ["metadata" ].get ("flags" , [])
167
- ],
168
- )
169
- bundles_to_upload .append ("startup.zip" )
170
-
171
139
# Build attachments bundle for collections which have the option set.
172
140
for changeset in all_changesets :
173
141
bid = changeset ["metadata" ]["bucket" ]
@@ -179,19 +147,27 @@ def build_bundles(event, context):
179
147
bundles_to_delete .append (attachments_bundle_filename )
180
148
if not BUILD_ALL :
181
149
continue
150
+ else :
151
+ print (f"{ bid } /{ cid } has attachments bundles enabled" )
182
152
153
+ existing_bundle_timestamp = get_modified_timestamp (
154
+ f"{ base_url } { DESTINATION_FOLDER } /{ bid } --{ cid } .zip"
155
+ )
156
+ print (f"'{ bid } --{ cid } .zip' was modified at { existing_bundle_timestamp } " )
157
+ print (f"Latest change on { bid } /{ cid } was at { changeset ["timestamp" ]} " )
183
158
if not BUILD_ALL and changeset ["timestamp" ] < existing_bundle_timestamp :
184
159
# Collection hasn't changed since last bundling.
160
+ print (f"{ bid } /{ cid } hasn't changed since last bundle." )
185
161
continue
186
162
187
163
# Skip bundle if no attachments found.
188
164
records = [r for r in changeset ["changes" ] if "attachment" in r ]
189
165
if not records :
190
- print ("%s/%s has no attachments" % ( bid , cid ) )
166
+ print (f" { bid } / { cid } has no attachments" )
191
167
bundles_to_delete .append (attachments_bundle_filename )
192
168
continue
193
169
194
- print ("%s/%s: %s records with attachments" % ( bid , cid , len ( records )) )
170
+ print (f" { bid } / { cid } { len ( records ) } records with attachments" )
195
171
196
172
# Skip bundle if total size is too big.
197
173
total_size_bytes = sum (r ["attachment" ]["size" ] for r in records )
@@ -211,6 +187,55 @@ def build_bundles(event, context):
211
187
)
212
188
bundles_to_upload .append (attachments_bundle_filename )
213
189
190
+ highest_timestamp = max (c ["timestamp" ] for c in all_changesets )
191
+ print (f"Latest server change was at { highest_timestamp } " )
192
+
193
+ existing_bundle_timestamp = get_modified_timestamp (
194
+ f"{ base_url } { DESTINATION_FOLDER } /changesets.zip"
195
+ )
196
+ if existing_bundle_timestamp is None :
197
+ print ("No previous 'changesets.zip' bundle found" ) # Should only happen once.
198
+ existing_bundle_timestamp = - 1
199
+ print (f"'changesets.zip' was published at { existing_bundle_timestamp } " )
200
+ if BUILD_ALL or (existing_bundle_timestamp < highest_timestamp ):
201
+ write_zip (
202
+ "changesets.zip" ,
203
+ [
204
+ (
205
+ "{metadata[bucket]}--{metadata[id]}.json" .format (** changeset ),
206
+ json .dumps (changeset ),
207
+ )
208
+ for changeset in all_changesets
209
+ ],
210
+ )
211
+ bundles_to_upload .append ("changesets.zip" )
212
+ else :
213
+ print ("Existing 'changesets.zip' bundle up-to-date. Nothing to do." )
214
+
215
+ # Build a bundle for collections that are marked with "startup" flag.
216
+ existing_bundle_timestamp = get_modified_timestamp (
217
+ f"{ base_url } { DESTINATION_FOLDER } /startup.zip"
218
+ )
219
+ if existing_bundle_timestamp is None :
220
+ print ("No previous 'startup.zip' bundle found" ) # Should only happen once.
221
+ existing_bundle_timestamp = - 1
222
+ print (f"'startup.zip' was published at { existing_bundle_timestamp } " )
223
+ if BUILD_ALL or existing_bundle_timestamp < highest_timestamp :
224
+ write_zip (
225
+ "startup.zip" ,
226
+ [
227
+ (
228
+ "{metadata[bucket]}--{metadata[id]}.json" .format (** changeset ),
229
+ json .dumps (changeset ),
230
+ )
231
+ for changeset in all_changesets
232
+ if "startup" in changeset ["metadata" ].get ("flags" , [])
233
+ ],
234
+ )
235
+ bundles_to_upload .append ("startup.zip" )
236
+ else :
237
+ print ("Existing 'startup.zip' bundle up-to-date. Nothing to do." )
238
+
214
239
if not SKIP_UPLOAD :
215
240
sync_cloud_storage (
216
241
STORAGE_BUCKET_NAME , DESTINATION_FOLDER , bundles_to_upload , bundles_to_delete
0 commit comments