forked from kernelci/kcidb
-
Notifications
You must be signed in to change notification settings - Fork 0
/
get-url-stats
executable file
·131 lines (108 loc) · 3.98 KB
/
get-url-stats
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
#!/usr/bin/env python3
"URL info caching script."
import csv
import json
import os
import sys
from urllib.parse import urlparse
import requests
OBJECT_SPECS = {
'checkouts': {
'patchset_files': [{'url': True}],
},
'builds': {
'input_files': [{'url': True}],
'output_files': [{'url': True}],
'config_url': True,
'log_url': True,
},
'tests': {
'output_files': [{'url': True}],
'log_url': True,
},
}
def extract_fields(spec, data):
"""
Extract values of fields from data according to a specification.
Args:
spec: The specification of fields to extract the values from.
data: The data to extract the field values from.
Returns:
An array of tuples, where each tuple contains:
* A tuple containing dictionary keys leading to the field.
* The field value.
"""
tuples = []
if spec is True:
return [(tuple(), data)]
if isinstance(spec, dict) and isinstance(data, dict):
for obj_type, obj_spec in spec.items():
obj = data.get(obj_type)
if obj is not None:
inner_tuples = extract_fields(obj_spec, obj)
# prefix the path from inner tuples with obj_type
for field_path, field_value in inner_tuples:
tuples.append(((obj_type,) + field_path, field_value))
elif isinstance(spec, list) and isinstance(data, list):
assert len(spec) == 1
obj_spec = spec[0]
for obj in data:
tuples += extract_fields(obj_spec, obj)
return tuples
def get_url_info(url):
"""Extract url info and returns fields value."""
status_code = 0
content_type = ''
size = ''
extension = os.path.splitext(urlparse(url).path)[1]
url_length = sys.getsizeof(url.encode('utf-8'))
# Some CI systems send links to index pages, instead of actual files,
# which is against the schema, and we don't really want to cache them.
# That's why we ignore URLs ending with "/"
if url.endswith('/'):
return (status_code, content_type, size, extension, url_length)
try:
response = requests.head(url, timeout=10)
except requests.exceptions.RequestException as error:
print(f"Error while trying to get URL info for {url}: {error!r}")
else:
status_code = response.status_code
if status_code == 200:
content_type = response.headers.get('Content-Type')
size = response.headers.get('Content-Length')
return (status_code, content_type, size, extension, url_length)
def main():
"""Extract URL info from JSON data and save it in a CSV file."""
# Read JSON file path from the terminal
json_file_path = sys.argv[1]
# Read JSON file
with open(json_file_path, 'r', encoding='utf-8') as json_file:
data = json.load(json_file)
# Prepare CSV file
csv_writer = csv.writer(sys.stdout)
csv_writer.writerow([
'Origin', 'Field Path', 'URL', 'Status Code',
'Content Type', 'File Extension', 'URL Length (Bytes)', 'Size'
])
url_path_set = set() # Track unique field path and URL combinations
# For each object type and spec
for obj_type, obj_spec in OBJECT_SPECS.items():
for obj in data.get(obj_type, []):
origin = obj.get("origin", "")
url_tuples = extract_fields(obj_spec, obj)
# Iterating through URL tuples
for keys, url in url_tuples:
path = '.'.join(str(key) for key in (obj_type,) + keys)
url_info = get_url_info(url)
# Check for unique field path and URL combination
if (path, url) in url_path_set:
continue
url_path_set.add((path, url))
# Save URL info in CSV file
csv_writer.writerow([
origin,
path, url, *url_info
])
sys.stdout.flush()
if __name__ == '__main__':
main()