commit
a39d9fa040
1 changed files with 101 additions and 0 deletions
@ -0,0 +1,101 @@ |
||||
#!/usr/bin/env python |
||||
# Copyright 2015 gRPC authors. |
||||
# |
||||
# Licensed under the Apache License, Version 2.0 (the "License"); |
||||
# you may not use this file except in compliance with the License. |
||||
# You may obtain a copy of the License at |
||||
# |
||||
# http://www.apache.org/licenses/LICENSE-2.0 |
||||
# |
||||
# Unless required by applicable law or agreed to in writing, software |
||||
# distributed under the License is distributed on an "AS IS" BASIS, |
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
||||
# See the License for the specific language governing permissions and |
||||
# limitations under the License. |
||||
|
||||
"""Detect new flakes introduced in the last 24h hours with respect to the |
||||
previous six days""" |
||||
|
||||
from __future__ import absolute_import |
||||
from __future__ import division |
||||
from __future__ import print_function |
||||
|
||||
import os |
||||
import sys |
||||
import logging |
||||
logging.basicConfig(format='%(asctime)s %(message)s') |
||||
|
||||
gcp_utils_dir = os.path.abspath( |
||||
os.path.join(os.path.dirname(__file__), '../gcp/utils')) |
||||
sys.path.append(gcp_utils_dir) |
||||
|
||||
import big_query_utils |
||||
|
||||
|
||||
def get_flaky_tests(days_lower_bound, days_upper_bound, limit=None): |
||||
""" period is one of "WEEK", "DAY", etc. |
||||
(see https://cloud.google.com/bigquery/docs/reference/standard-sql/functions-and-operators#date_add). """ |
||||
|
||||
bq = big_query_utils.create_big_query() |
||||
query = """ |
||||
SELECT |
||||
filtered_test_name, |
||||
FIRST(timestamp), |
||||
FIRST(build_url), |
||||
FROM ( |
||||
SELECT |
||||
REGEXP_REPLACE(test_name, r'/\d+', '') AS filtered_test_name, |
||||
result, |
||||
build_url, |
||||
timestamp |
||||
FROM |
||||
[grpc-testing:jenkins_test_results.aggregate_results] |
||||
WHERE |
||||
timestamp >= DATE_ADD(CURRENT_DATE(), {days_lower_bound}, "DAY") |
||||
AND timestamp <= DATE_ADD(CURRENT_DATE(), {days_upper_bound}, "DAY") |
||||
AND NOT REGEXP_MATCH(job_name, '.*portability.*')) |
||||
GROUP BY |
||||
filtered_test_name, |
||||
timestamp, |
||||
build_url |
||||
HAVING |
||||
SUM(result != 'PASSED' |
||||
AND result != 'SKIPPED') > 0 |
||||
ORDER BY |
||||
timestamp ASC |
||||
""".format(days_lower_bound=days_lower_bound, days_upper_bound=days_upper_bound) |
||||
if limit: |
||||
query += '\n LIMIT {}'.format(limit) |
||||
query_job = big_query_utils.sync_query_job(bq, 'grpc-testing', query) |
||||
page = bq.jobs().getQueryResults( |
||||
pageToken=None, **query_job['jobReference']).execute(num_retries=3) |
||||
testname_to_ts_url_pair = {row['f'][0]['v']: (row['f'][1]['v'], row['f'][2]['v']) for row in page['rows']} |
||||
return testname_to_ts_url_pair |
||||
|
||||
|
||||
def get_new_flakes(): |
||||
last_week_sans_yesterday = get_flaky_tests(-7, -1) |
||||
last_24 = get_flaky_tests(-1, +1) |
||||
last_week_sans_yesterday_names = set(last_week_sans_yesterday.keys()) |
||||
last_24_names = set(last_24.keys()) |
||||
logging.debug('|last_week_sans_yesterday| =', len(last_week_sans_yesterday_names)) |
||||
logging.debug('|last_24_names| =', len(last_24_names)) |
||||
new_flakes = last_24_names - last_week_sans_yesterday_names |
||||
logging.debug('|new_flakes| = ', len(new_flakes)) |
||||
return {k: last_24[k] for k in new_flakes} |
||||
|
||||
|
||||
def main(): |
||||
import datetime |
||||
new_flakes = get_new_flakes() |
||||
if new_flakes: |
||||
print("Found {} new flakes:".format(len(new_flakes))) |
||||
for k, v in new_flakes.items(): |
||||
ts = int(float(v[0])) |
||||
url = v[1] |
||||
human_ts = datetime.datetime.utcfromtimestamp(ts).strftime('%Y-%m-%d %H:%M:%S UTC') |
||||
print("Test: {}, Timestamp: {}, URL: {}\n".format(k, human_ts, url)) |
||||
|
||||
|
||||
if __name__ == '__main__': |
||||
main() |
Loading…
Reference in new issue