Revision 9bee4afa
Added by koszko over 1 year ago
.gitmodules | ||
---|---|---|
1 | 1 |
[submodule "schemas"] |
2 |
path = schemas |
|
2 |
path = schemas/1.x |
|
3 |
url = ../hydrilla-json-schemas/ |
|
4 |
[submodule "hydrilla-json-schemas-2.x"] |
|
5 |
path = schemas/2.x |
|
3 | 6 |
url = ../hydrilla-json-schemas/ |
common/jsonschema.js | ||
---|---|---|
67 | 67 |
#EXPORT validate |
68 | 68 |
|
69 | 69 |
const haketilo_schemas = [ |
70 |
#INCLUDE schemas/api_query_result-1.0.1.schema.json |
|
70 |
/* 1.x Hydrilla JSON schema series */ |
|
71 |
#INCLUDE schemas/1.x/api_query_result-1.0.1.schema.json |
|
71 | 72 |
, |
72 |
#INCLUDE schemas/api_mapping_description-1.0.1.schema.json |
|
73 |
#INCLUDE schemas/1.x/api_mapping_description-1.0.1.schema.json
|
|
73 | 74 |
, |
74 |
#INCLUDE schemas/api_resource_description-1.0.1.schema.json |
|
75 |
#INCLUDE schemas/1.x/api_resource_description-1.0.1.schema.json
|
|
75 | 76 |
, |
76 |
#INCLUDE schemas/common_definitions-1.0.1.schema.json |
|
77 |
#INCLUDE schemas/1.x/common_definitions-1.0.1.schema.json |
|
78 |
, |
|
79 |
/* 2.x Hydrilla JSON schema series */ |
|
80 |
#INCLUDE schemas/2.x/api_query_result-2.schema.json |
|
81 |
, |
|
82 |
#INCLUDE schemas/2.x/api_mapping_description-2.schema.json |
|
83 |
, |
|
84 |
#INCLUDE schemas/2.x/api_resource_description-2.schema.json |
|
85 |
, |
|
86 |
#INCLUDE schemas/2.x/common_definitions-2.schema.json |
|
77 | 87 |
].reduce((ac, s) => Object.assign(ac, {[s.$id]: s}), {}); |
88 |
|
|
89 |
const name_base_re = "(?<name_base>[^/]*)"; |
|
90 |
const major_number_re = "(?<major>[1-9][0-9]*)"; |
|
91 |
const minor_number_re = "(?:[1-9][0-9]*|0)"; |
|
92 |
const numbers_rest_re = `(?:\\.${minor_number_re})*`; |
|
93 |
const version_re = `(?<ver>${major_number_re}${numbers_rest_re})`; |
|
94 |
const schema_name_re = `${name_base_re}-${version_re}\\.schema\\.json`; |
|
95 |
|
|
96 |
const haketilo_schema_name_regex = new RegExp(schema_name_re); |
|
97 |
|
|
98 |
for (const [$id, schema] of [...Object.entries(haketilo_schemas)]) { |
|
99 |
const match = haketilo_schema_name_regex.exec($id); |
|
100 |
const schema_name = |
|
101 |
`${match.groups.name_base}-${match.groups.major}.schema.json`; |
|
102 |
haketilo_schemas[schema_name] = schema; |
|
103 |
} |
|
104 |
|
|
78 | 105 |
#EXPORT haketilo_schemas |
106 |
#EXPORT haketilo_schema_name_regex |
|
79 | 107 |
|
80 | 108 |
const haketilo_validator = new Validator(); |
81 | 109 |
Object.values(haketilo_schemas) |
html/install.js | ||
---|---|---|
49 | 49 |
#FROM html/DOM_helpers.js IMPORT clone_template, Showable |
50 | 50 |
#FROM common/entities.js IMPORT item_id_string, version_string, get_files |
51 | 51 |
#FROM common/misc.js IMPORT sha256_async AS compute_sha256 |
52 |
#FROM common/jsonschema.js IMPORT haketilo_validator, haketilo_schemas |
|
52 |
#FROM common/jsonschema.js IMPORT haketilo_validator, haketilo_schemas, \ |
|
53 |
haketilo_schema_name_regex |
|
53 | 54 |
|
54 | 55 |
#FROM html/repo_query_cacher_client.js IMPORT indirect_fetch |
55 | 56 |
|
... | ... | |
203 | 204 |
|
204 | 205 |
const captype = item_type[0].toUpperCase() + item_type.substring(1); |
205 | 206 |
|
206 |
const $id = |
|
207 |
`https://hydrilla.koszko.org/schemas/api_${item_type}_description-1.0.1.schema.json`; |
|
208 |
const schema = haketilo_schemas[$id]; |
|
209 |
const result = haketilo_validator.validate(json, schema); |
|
210 |
if (result.errors.length > 0) { |
|
211 |
const reg = new RegExp(schema.allOf[2].properties.$schema.pattern); |
|
212 |
if (json.$schema && !reg.test(json.$schema)) { |
|
207 |
const nonconforming_format_error_msg = |
|
208 |
`${captype} ${item_id_string(id, ver)} was served using a nonconforming response format.`; |
|
209 |
|
|
210 |
try { |
|
211 |
const match = haketilo_schema_name_regex.exec(json.$schema); |
|
212 |
var major_schema_version = match.groups.major; |
|
213 |
|
|
214 |
if (!["1", "2"].includes(major_schema_version)) { |
|
213 | 215 |
const msg = `${captype} ${item_id_string(id, ver)} was served using unsupported Hydrilla API version. You might need to update Haketilo.`; |
214 |
return work.err(result.errors, msg);
|
|
216 |
return work.err(null, msg);
|
|
215 | 217 |
} |
216 |
|
|
217 |
const msg = `${captype} ${item_id_string(id, ver)} was served using a nonconforming response format.`; |
|
218 |
return work.err(result.errors, msg); |
|
218 |
} catch(e) { |
|
219 |
return work.err(e, nonconforming_format_error_msg); |
|
219 | 220 |
} |
220 | 221 |
|
222 |
const schema_name = `api_${item_type}_description-${major_schema_version}.schema.json`; |
|
223 |
|
|
224 |
const schema = haketilo_schemas[schema_name]; |
|
225 |
const result = haketilo_validator.validate(json, schema); |
|
226 |
if (result.errors.length > 0) |
|
227 |
return work.err(result.errors, nonconforming_format_error_msg); |
|
228 |
|
|
221 | 229 |
const scripts = item_type === "resource" && json.scripts; |
222 | 230 |
const files = json.source_copyright.concat(scripts || []); |
223 | 231 |
|
... | ... | |
229 | 237 |
process_item(work, "resource", res_ref.identifier); |
230 | 238 |
} |
231 | 239 |
|
240 |
if (major_schema_version >= 2) { |
|
241 |
for (const map_ref of (json.required_mappings || [])) |
|
242 |
process_item(work, "mapping", map_ref.identifier); |
|
243 |
} |
|
244 |
|
|
232 | 245 |
/* |
233 | 246 |
* At this point we already have JSON definition of the item and we |
234 | 247 |
* triggered processing of its dependencies. We now have to verify if |
schemas | ||
---|---|---|
1 |
Subproject commit 09634f3446866f712a022327683b1149d8f46bf0 |
schemas/1.x | ||
---|---|---|
1 |
Subproject commit 09634f3446866f712a022327683b1149d8f46bf0 |
schemas/2.x | ||
---|---|---|
1 |
Subproject commit 7206db45f277c10c34d1b7ed9bd35343ac742d30 |
test/haketilo_test/unit/test_install.py | ||
---|---|---|
57 | 57 |
|
58 | 58 |
@pytest.mark.ext_data(install_ext_data) |
59 | 59 |
@pytest.mark.usefixtures('webextension') |
60 |
@pytest.mark.parametrize('complex_variant', [False, True]) |
|
61 |
def test_install_normal_usage(driver, execute_in_page, complex_variant): |
|
60 |
@pytest.mark.parametrize('variant', [{ |
|
61 |
# The resource/mapping others depend on. |
|
62 |
'root_resource_id': f'resource-abcd-defg-ghij', |
|
63 |
'root_mapping_id': f'mapping-abcd-defg-ghij', |
|
64 |
# Those ids are used to check the alphabetical ordering. |
|
65 |
'item_ids': [f'resource-{letters}' for letters in ( |
|
66 |
'a', 'abcd', 'abcd-defg-ghij', 'b', 'c', |
|
67 |
'd', 'defg', 'e', 'f', |
|
68 |
'g', 'ghij', 'h', 'i', 'j' |
|
69 |
)], |
|
70 |
'files_count': 9 |
|
71 |
}, { |
|
72 |
'root_resource_id': 'resource-a', |
|
73 |
'root_mapping_id': 'mapping-a', |
|
74 |
'item_ids': ['resource-a'], |
|
75 |
'files_count': 0 |
|
76 |
}, { |
|
77 |
'root_resource_id': 'resource-a-w-required-mapping-v1', |
|
78 |
'root_mapping_id': 'mapping-a-w-required-mapping-v1', |
|
79 |
'item_ids': ['resource-a-w-required-mapping-v1'], |
|
80 |
'files_count': 1 |
|
81 |
}, { |
|
82 |
'root_resource_id': 'resource-a-w-required-mapping-v2', |
|
83 |
'root_mapping_id': 'mapping-a-w-required-mapping-v2', |
|
84 |
'item_ids': [ |
|
85 |
'mapping-a', |
|
86 |
'resource-a', |
|
87 |
'resource-a-w-required-mapping-v2' |
|
88 |
], |
|
89 |
'files_count': 1 |
|
90 |
}]) |
|
91 |
def test_install_normal_usage(driver, execute_in_page, variant): |
|
62 | 92 |
""" |
63 | 93 |
Test of the normal package installation procedure with one mapping and, |
64 | 94 |
depending on parameter, one or many resources. |
... | ... | |
67 | 97 |
|
68 | 98 |
assert execute_in_page('returnval(shw());') == [[], False] |
69 | 99 |
|
70 |
if complex_variant: |
|
71 |
# The resource/mapping others depend on. |
|
72 |
root_id = 'abcd-defg-ghij' |
|
73 |
root_resource_id = f'resource-{root_id}' |
|
74 |
root_mapping_id = f'mapping-{root_id}' |
|
75 |
# Those ids are used to check the alphabetical ordering. |
|
76 |
resource_ids = [f'resource-{letters}' for letters in ( |
|
77 |
'a', 'abcd', root_id, 'b', 'c', |
|
78 |
'd', 'defg', 'e', 'f', |
|
79 |
'g', 'ghij', 'h', 'i', 'j' |
|
80 |
)] |
|
81 |
files_count = 9 |
|
82 |
else: |
|
83 |
root_resource_id = f'resource-a' |
|
84 |
root_mapping_id = f'mapping-a' |
|
85 |
resource_ids = [root_resource_id] |
|
86 |
files_count = 0 |
|
87 |
|
|
88 | 100 |
# Preview the installation of a resource, show resource's details, close |
89 | 101 |
# the details and cancel installation. |
90 | 102 |
execute_in_page('returnval(install_view.show(...arguments));', |
91 |
'https://hydril.la/', 'resource', root_resource_id) |
|
103 |
'https://hydril.la/', 'resource', |
|
104 |
variant['root_resource_id']) |
|
92 | 105 |
|
93 | 106 |
assert execute_in_page('returnval(shw());') == [['show'], True] |
94 |
assert f'{root_resource_id}-2021.11.11-1'\
|
|
107 |
assert f'{variant["root_resource_id"]}-2021.11.11-1'\
|
|
95 | 108 |
in containers['install_preview'].text |
96 | 109 |
assert_container_displayed('install_preview') |
97 | 110 |
|
98 | 111 |
entries = execute_in_page('returnval(ets().map(e => e.main_li.innerText));') |
99 |
assert len(entries) == len(resource_ids) |
|
112 |
assert len(entries) == len(variant['item_ids']) |
|
113 |
resource_idx = variant['item_ids'].index(variant['root_resource_id']) |
|
100 | 114 |
# Verify alphabetical ordering. |
101 |
assert all([id in text for id, text in zip(resource_ids, entries)]) |
|
115 |
assert all([id in text for id, text in |
|
116 |
zip(variant['item_ids'], entries)]) |
|
102 | 117 |
|
103 |
assert not execute_in_page('returnval(ets()[0].old_ver);').is_displayed() |
|
104 |
execute_in_page('returnval(ets()[0].details_but);').click() |
|
118 |
assert not execute_in_page(f'returnval(ets()[{resource_idx}].old_ver);')\ |
|
119 |
.is_displayed() |
|
120 |
execute_in_page(f'returnval(ets()[{resource_idx}].details_but);').click() |
|
105 | 121 |
assert 'resource-a' in containers['resource_preview_container'].text |
106 | 122 |
assert_container_displayed('resource_preview_container') |
107 | 123 |
|
... | ... | |
116 | 132 |
# details, close the details and commit the installation. |
117 | 133 |
execute_in_page('returnval(install_view.show(...arguments));', |
118 | 134 |
'https://hydril.la/', 'mapping', |
119 |
root_mapping_id, [2022, 5, 10])
|
|
135 |
variant['root_mapping_id'], [2022, 5, 10])
|
|
120 | 136 |
|
121 | 137 |
assert execute_in_page('returnval(shw(2));') == [['show'], True] |
122 | 138 |
assert_container_displayed('install_preview') |
123 | 139 |
|
124 | 140 |
entries = execute_in_page('returnval(ets().map(e => e.main_li.innerText));') |
125 |
assert len(entries) == len(resource_ids) + 1 |
|
126 |
assert f'{root_mapping_id}-2022.5.10' in entries[0] |
|
141 |
assert len(entries) == len(variant['item_ids']) + 1 |
|
142 |
|
|
143 |
all_item_ids = sorted([*variant['item_ids'], variant['root_mapping_id']]) |
|
144 |
mapping_idx = all_item_ids.index(variant["root_mapping_id"]) |
|
127 | 145 |
# Verify alphabetical ordering. |
128 |
assert all([id in text for id, text in zip(resource_ids, entries[1:])])
|
|
146 |
assert all([id in text for id, text in zip(all_item_ids, entries)])
|
|
129 | 147 |
|
130 |
assert not execute_in_page('returnval(ets()[0].old_ver);').is_displayed() |
|
131 |
execute_in_page('returnval(ets()[0].details_but);').click() |
|
132 |
assert root_mapping_id in containers['mapping_preview_container'].text |
|
148 |
assert not execute_in_page(f'returnval(ets()[{mapping_idx}].old_ver);')\ |
|
149 |
.is_displayed() |
|
150 |
execute_in_page(f'returnval(ets()[{mapping_idx}].details_but);').click() |
|
151 |
assert variant['root_mapping_id'] in \ |
|
152 |
containers['mapping_preview_container'].text |
|
133 | 153 |
assert_container_displayed('mapping_preview_container') |
134 | 154 |
|
135 | 155 |
execute_in_page('returnval(install_view.mapping_back_but);').click() |
... | ... | |
145 | 165 |
|
146 | 166 |
# Verify the install |
147 | 167 |
db_contents = get_db_contents(execute_in_page) |
148 |
for item_type, ids in \ |
|
149 |
[('mapping', {root_mapping_id}), ('resource', set(resource_ids))]: |
|
168 |
all_map_ids = {id for id in all_item_ids if id.startswith('mapping')} |
|
169 |
all_res_ids = {id for id in all_item_ids if id.startswith('resource')} |
|
170 |
for item_type, ids in [ |
|
171 |
('mapping', all_map_ids), |
|
172 |
('resource', all_res_ids) |
|
173 |
]: |
|
150 | 174 |
assert set([it['identifier'] for it in db_contents[item_type]]) == ids |
151 | 175 |
|
152 |
assert all([len(db_contents[store]) == files_count
|
|
176 |
assert all([len(db_contents[store]) == variant['files_count']
|
|
153 | 177 |
for store in ('file', 'file_uses')]) |
154 | 178 |
|
155 | 179 |
# Update the installed mapping to a newer version. |
156 | 180 |
execute_in_page('returnval(install_view.show(...arguments));', |
157 |
'https://hydril.la/', 'mapping', root_mapping_id)
|
|
181 |
'https://hydril.la/', 'mapping', variant['root_mapping_id'])
|
|
158 | 182 |
assert execute_in_page('returnval(shw(4));') == [['show'], True] |
159 | 183 |
# resources are already in the newest versions, hence they should not appear |
160 | 184 |
# in the install preview list. |
... | ... | |
171 | 195 |
|
172 | 196 |
# Verify the newer version install. |
173 | 197 |
old_db_contents, db_contents = db_contents, get_db_contents(execute_in_page) |
174 |
old_db_contents['mapping'][0]['version'][-1] += 1 |
|
175 |
assert db_contents['mapping'] == old_db_contents['mapping'] |
|
198 |
|
|
199 |
old_root_mapping = [m for m in old_db_contents['mapping'] |
|
200 |
if m['identifier'] == variant['root_mapping_id']][0] |
|
201 |
old_root_mapping['version'][-1] += 1 |
|
202 |
|
|
203 |
new_root_mapping = [m for m in db_contents['mapping'] |
|
204 |
if m['identifier'] == variant['root_mapping_id']][0] |
|
205 |
|
|
206 |
assert old_root_mapping == new_root_mapping |
|
176 | 207 |
|
177 | 208 |
# All items are up to date - verify dialog is instead shown in this case. |
178 | 209 |
execute_in_page('install_view.show(...arguments);', |
179 |
'https://hydril.la/', 'mapping', root_mapping_id)
|
|
210 |
'https://hydril.la/', 'mapping', variant['root_mapping_id'])
|
|
180 | 211 |
|
181 | 212 |
fetched = lambda d: 'Fetching ' not in containers['dialog_container'].text |
182 | 213 |
WebDriverWait(driver, 10).until(fetched) |
test/haketilo_test/world_wide_library.py | ||
---|---|---|
33 | 33 |
from threading import Lock |
34 | 34 |
from uuid import uuid4 |
35 | 35 |
import json |
36 |
import functools as ft |
|
37 |
import operator as op |
|
36 | 38 |
|
37 | 39 |
from .misc_constants import here |
38 | 40 |
from .unit.utils import * # sample repo data |
... | ... | |
114 | 116 |
for i in range(9)] |
115 | 117 |
sample_hashes = [sha256(c.encode()).digest().hex() for c in sample_contents] |
116 | 118 |
|
117 |
file_url = lambda hashed: f'https://hydril.la/file/sha256/{hashed}'
|
|
119 |
file_url = ft.partial(op.concat, 'https://hydril.la/file/sha256/')
|
|
118 | 120 |
|
119 | 121 |
sample_files_catalog = dict([(file_url(h), make_handler(c)) |
120 | 122 |
for h, c in zip(sample_hashes, sample_contents)]) |
... | ... | |
144 | 146 |
'dependencies': [] |
145 | 147 |
}) |
146 | 148 |
|
149 |
# The one below will generate items with schema still at version 1, so required |
|
150 |
# mappings will be ignored. |
|
151 |
sample_resource_templates.append({ |
|
152 |
'id_suffix': 'a-w-required-mapping-v1', |
|
153 |
'files_count': 1, |
|
154 |
'dependencies': [], |
|
155 |
'required_mappings': [{'identifier': 'mapping-a'}] |
|
156 |
}) |
|
157 |
|
|
158 |
sample_resource_templates.append({ |
|
159 |
'id_suffix': 'a-w-required-mapping-v2', |
|
160 |
'files_count': 1, |
|
161 |
'dependencies': [], |
|
162 |
'required_mappings': [{'identifier': 'mapping-a'}], |
|
163 |
'schema_ver': '2' |
|
164 |
}) |
|
165 |
|
|
147 | 166 |
sample_resources_catalog = {} |
148 | 167 |
sample_mappings_catalog = {} |
149 | 168 |
sample_queries = {} |
150 | 169 |
|
151 | 170 |
for srt in sample_resource_templates: |
152 | 171 |
resource = make_sample_resource() |
153 |
resource['identifier'] = f'resource-{srt["id_suffix"]}'
|
|
154 |
resource['long_name'] = resource['identifier'].upper()
|
|
155 |
resource['uuid'] = str(uuid4())
|
|
156 |
resource['dependencies'] = srt['dependencies']
|
|
157 |
resource['source_copyright'] = []
|
|
158 |
resource['scripts'] = []
|
|
172 |
resource['identifier'] = f'resource-{srt["id_suffix"]}' |
|
173 |
resource['long_name'] = resource['identifier'].upper() |
|
174 |
resource['uuid'] = str(uuid4()) |
|
175 |
resource['dependencies'] = srt['dependencies'] |
|
176 |
resource['source_copyright'] = [] |
|
177 |
resource['scripts'] = [] |
|
159 | 178 |
for i in range(srt['files_count']): |
160 | 179 |
file_ref = {'file': f'file_{i}', 'sha256': sample_hashes[i]} |
161 | 180 |
resource[('source_copyright', 'scripts')[i & 1]].append(file_ref) |
... | ... | |
191 | 210 |
|
192 | 211 |
mapping['payloads'] = payloads |
193 | 212 |
|
213 |
for item in resource, mapping: |
|
214 |
if 'required_mappings' in srt: |
|
215 |
item['required_mappings'] = srt['required_mappings'] |
|
216 |
if 'schema_ver' in srt: |
|
217 |
item['$schema'] = item['$schema'].replace('1', srt['schema_ver']) |
|
218 |
|
|
194 | 219 |
for item, versions, catalog in [ |
195 | 220 |
(resource, resource_versions, sample_resources_catalog), |
196 | 221 |
(mapping, mapping_versions, sample_mappings_catalog) |
Also available in: Unified diff
support schema v2 and dependencies on mappings