Skip to content
This repository was archived by the owner on Feb 26, 2025. It is now read-only.

Commit da7af1c

Browse files
authored
Improve build_bundles idempotence (#1488)
1 parent b584058 commit da7af1c

File tree

2 files changed

+81
-55
lines changed

2 files changed

+81
-55
lines changed

commands/build_bundles.py

+59-34
Original file line numberDiff line numberDiff line change
@@ -127,19 +127,7 @@ def build_bundles(event, context):
127127

128128
base_url = client.server_info()["capabilities"]["attachments"]["base_url"]
129129

130-
existing_bundle_timestamp = get_modified_timestamp(
131-
f"{base_url}{DESTINATION_FOLDER}/changesets.zip"
132-
)
133-
if existing_bundle_timestamp is None:
134-
print("No previous bundle found") # Should only happen once.
135-
existing_bundle_timestamp = -1
136-
137130
all_changesets = fetch_all_changesets(client)
138-
highest_timestamp = max(c["timestamp"] for c in all_changesets)
139-
140-
if existing_bundle_timestamp >= highest_timestamp:
141-
print("Existing bundle up-to-date. Nothing to do.")
142-
return
143131

144132
# Build all archives in temp directory.
145133
tmp_dir = tempfile.mkdtemp()
@@ -148,26 +136,6 @@ def build_bundles(event, context):
148136
bundles_to_upload = []
149137
bundles_to_delete = []
150138

151-
write_zip(
152-
"changesets.zip",
153-
[
154-
("{metadata[bucket]}--{metadata[id]}.json".format(**changeset), json.dumps(changeset))
155-
for changeset in all_changesets
156-
],
157-
)
158-
bundles_to_upload.append("changesets.zip")
159-
160-
# Build a bundle for collections that are marked with "startup" flag.
161-
write_zip(
162-
"startup.zip",
163-
[
164-
("{metadata[bucket]}--{metadata[id]}.json".format(**changeset), json.dumps(changeset))
165-
for changeset in all_changesets
166-
if "startup" in changeset["metadata"].get("flags", [])
167-
],
168-
)
169-
bundles_to_upload.append("startup.zip")
170-
171139
# Build attachments bundle for collections which have the option set.
172140
for changeset in all_changesets:
173141
bid = changeset["metadata"]["bucket"]
@@ -179,19 +147,27 @@ def build_bundles(event, context):
179147
bundles_to_delete.append(attachments_bundle_filename)
180148
if not BUILD_ALL:
181149
continue
150+
else:
151+
print(f"{bid}/{cid} has attachments bundles enabled")
182152

153+
existing_bundle_timestamp = get_modified_timestamp(
154+
f"{base_url}{DESTINATION_FOLDER}/{bid}--{cid}.zip"
155+
)
156+
print(f"'{bid}--{cid}.zip' was modified at {existing_bundle_timestamp}")
157+
print(f"Latest change on {bid}/{cid} was at {changeset["timestamp"]}")
183158
if not BUILD_ALL and changeset["timestamp"] < existing_bundle_timestamp:
184159
# Collection hasn't changed since last bundling.
160+
print(f"{bid}/{cid} hasn't changed since last bundle.")
185161
continue
186162

187163
# Skip bundle if no attachments found.
188164
records = [r for r in changeset["changes"] if "attachment" in r]
189165
if not records:
190-
print("%s/%s has no attachments" % (bid, cid))
166+
print(f"{bid}/{cid} has no attachments")
191167
bundles_to_delete.append(attachments_bundle_filename)
192168
continue
193169

194-
print("%s/%s: %s records with attachments" % (bid, cid, len(records)))
170+
print(f"{bid}/{cid} {len(records)} records with attachments")
195171

196172
# Skip bundle if total size is too big.
197173
total_size_bytes = sum(r["attachment"]["size"] for r in records)
@@ -211,6 +187,55 @@ def build_bundles(event, context):
211187
)
212188
bundles_to_upload.append(attachments_bundle_filename)
213189

190+
highest_timestamp = max(c["timestamp"] for c in all_changesets)
191+
print(f"Latest server change was at {highest_timestamp}")
192+
193+
existing_bundle_timestamp = get_modified_timestamp(
194+
f"{base_url}{DESTINATION_FOLDER}/changesets.zip"
195+
)
196+
if existing_bundle_timestamp is None:
197+
print("No previous 'changesets.zip' bundle found") # Should only happen once.
198+
existing_bundle_timestamp = -1
199+
print(f"'changesets.zip' was published at {existing_bundle_timestamp}")
200+
if BUILD_ALL or (existing_bundle_timestamp < highest_timestamp):
201+
write_zip(
202+
"changesets.zip",
203+
[
204+
(
205+
"{metadata[bucket]}--{metadata[id]}.json".format(**changeset),
206+
json.dumps(changeset),
207+
)
208+
for changeset in all_changesets
209+
],
210+
)
211+
bundles_to_upload.append("changesets.zip")
212+
else:
213+
print("Existing 'changesets.zip' bundle up-to-date. Nothing to do.")
214+
215+
# Build a bundle for collections that are marked with "startup" flag.
216+
existing_bundle_timestamp = get_modified_timestamp(
217+
f"{base_url}{DESTINATION_FOLDER}/startup.zip"
218+
)
219+
if existing_bundle_timestamp is None:
220+
print("No previous 'startup.zip' bundle found") # Should only happen once.
221+
existing_bundle_timestamp = -1
222+
print(f"'startup.zip' was published at {existing_bundle_timestamp}")
223+
if BUILD_ALL or existing_bundle_timestamp < highest_timestamp:
224+
write_zip(
225+
"startup.zip",
226+
[
227+
(
228+
"{metadata[bucket]}--{metadata[id]}.json".format(**changeset),
229+
json.dumps(changeset),
230+
)
231+
for changeset in all_changesets
232+
if "startup" in changeset["metadata"].get("flags", [])
233+
],
234+
)
235+
bundles_to_upload.append("startup.zip")
236+
else:
237+
print("Existing 'startup.zip' bundle up-to-date. Nothing to do.")
238+
214239
if not SKIP_UPLOAD:
215240
sync_cloud_storage(
216241
STORAGE_BUCKET_NAME, DESTINATION_FOLDER, bundles_to_upload, bundles_to_delete

tests/test_build_bundles.py

+22-21
Original file line numberDiff line numberDiff line change
@@ -148,13 +148,14 @@ def test_build_bundles(mock_fetch_all_changesets, mock_write_zip, mock_sync_clou
148148
)
149149
responses.add(responses.GET, f"{server_url}/attachments/file.jpg", body=b"jpeg_content")
150150

151-
responses.add(
152-
responses.GET,
153-
f"{server_url}/attachments/bundles/changesets.zip",
154-
headers={
155-
"Last-Modified": "Wed, 03 Jul 2024 11:04:48 GMT" # 1720004688000
156-
},
157-
)
151+
for bundle in ["changesets", "startup"] + [f"bucket{i}--collection{i}" for i in range(5)]:
152+
responses.add(
153+
responses.GET,
154+
f"{server_url}/attachments/bundles/{bundle}.zip",
155+
headers={
156+
"Last-Modified": "Wed, 03 Jul 2024 11:04:48 GMT" # 1720004688000
157+
},
158+
)
158159

159160
mock_fetch_all_changesets.return_value = [
160161
{ # collection hasn't changed since last bundling
@@ -167,7 +168,7 @@ def test_build_bundles(mock_fetch_all_changesets, mock_write_zip, mock_sync_clou
167168
},
168169
{
169170
"changes": [
170-
{"id": "record1", "attachment": {"location": "file.jpg", "size": 10}},
171+
{"id": "record1", "attachment": {"location": "file.jpg", "size": 10000000}},
171172
{"id": "record2"},
172173
],
173174
"metadata": {"id": "collection1", "bucket": "bucket1", "attachment": {"bundle": True}},
@@ -209,8 +210,16 @@ def test_build_bundles(mock_fetch_all_changesets, mock_write_zip, mock_sync_clou
209210
) # changesets.zip, startup.zip, and only one for the attachments
210211
calls = mock_write_zip.call_args_list
211212

212-
# Assert the first call (changesets.zip)
213-
changesets_zip_path, changesets_zip_files = calls[0][0]
213+
# Assert the first call (attachments zip)
214+
attachments_zip_path, attachments_zip_files = calls[0][0]
215+
assert attachments_zip_path == "bucket1--collection1.zip"
216+
assert len(attachments_zip_files) == 2
217+
assert attachments_zip_files[0][0] == "record1.meta.json"
218+
assert attachments_zip_files[1][0] == "record1"
219+
assert attachments_zip_files[1][1] == b"jpeg_content"
220+
221+
# Assert the second call (changesets.zip)
222+
changesets_zip_path, changesets_zip_files = calls[1][0]
214223
assert changesets_zip_path == "changesets.zip"
215224
assert len(changesets_zip_files) == 6
216225
assert changesets_zip_files[0][0] == "bucket0--collection0.json"
@@ -220,27 +229,19 @@ def test_build_bundles(mock_fetch_all_changesets, mock_write_zip, mock_sync_clou
220229
assert changesets_zip_files[4][0] == "bucket4--collection4.json"
221230
assert changesets_zip_files[5][0] == "bucket5--collection5.json"
222231

223-
# Assert the second call (startup.zip)
224-
startup_zip_path, startup_zip_files = calls[1][0]
232+
# Assert the third call (startup.zip)
233+
startup_zip_path, startup_zip_files = calls[2][0]
225234
assert startup_zip_path == "startup.zip"
226235
assert len(startup_zip_files) == 1
227236
assert startup_zip_files[0][0] == "bucket5--collection5.json"
228237

229-
# Assert the third call (attachments zip)
230-
attachments_zip_path, attachments_zip_files = calls[2][0]
231-
assert attachments_zip_path == "bucket1--collection1.zip"
232-
assert len(attachments_zip_files) == 2
233-
assert attachments_zip_files[0][0] == "record1.meta.json"
234-
assert attachments_zip_files[1][0] == "record1"
235-
assert attachments_zip_files[1][1] == b"jpeg_content"
236-
237238
mock_sync_cloud_storage.assert_called_once_with(
238239
"remote-settings-test-local-attachments",
239240
"bundles",
240241
[
242+
"bucket1--collection1.zip",
241243
"changesets.zip",
242244
"startup.zip",
243-
"bucket1--collection1.zip",
244245
],
245246
[
246247
"bucket2--collection2.zip",

0 commit comments

Comments
 (0)