|
14 | 14 | # the distance between commits to include in the output |
15 | 15 | day_distance = 1 |
16 | 16 |
|
17 | | -# the directory where codeql is. This is the directory where we change the SHAs |
18 | | -working_dir = sys.argv[1] |
19 | | - |
20 | | -lang = "java" |
21 | | -db = "empty-java" |
22 | | -ql_output = "output-java.csv" |
23 | | -csv_output = "timeseries-java.csv" |
24 | | - |
25 | 17 |
|
26 | 18 | def get_str_output(arr): |
27 | 19 | r = subprocess.check_output(arr) |
@@ -49,66 +41,82 @@ def get_previous_sha(sha, date): |
49 | 41 | return (parent_sha, parent_date) |
50 | 42 |
|
51 | 43 |
|
52 | | -def get_stats(): |
53 | | - if os.path.isdir(db): |
54 | | - shutil.rmtree(db) |
55 | | - utils.create_empty_database(lang, ".java", db) |
56 | | - utils.run_codeql_query( |
57 | | - "java/ql/src/meta/frameworks/Coverage.ql", db, ql_output) |
58 | | - shutil.rmtree(db) |
59 | | - |
60 | | - sources = 0 |
61 | | - sinks = 0 |
62 | | - summaries = 0 |
63 | | - |
64 | | - with open(ql_output) as csvfile: |
65 | | - reader = csv.reader(csvfile) |
66 | | - for row in reader: |
67 | | - # row: "android.util",1,"remote","source",16 |
68 | | - if row[3] == "source": |
69 | | - sources += int(row[4]) |
70 | | - if row[3] == "sink": |
71 | | - sinks += int(row[4]) |
72 | | - if row[3] == "summary": |
73 | | - summaries += int(row[4]) |
74 | | - |
75 | | - os.remove(ql_output) |
76 | | - |
77 | | - return (sources, sinks, summaries) |
78 | | - |
79 | | - |
80 | | -with open(csv_output, 'w', newline='') as csvfile: |
81 | | - csvwriter = csv.writer(csvfile) |
82 | | - csvwriter.writerow(["SHA", "Date", "Sources", "Sinks", "Summaries"]) |
83 | | - |
84 | | - os.chdir(working_dir) |
| 44 | +def get_stats(lang, query): |
| 45 | + try: |
| 46 | + db = "empty_" + lang |
| 47 | + ql_output = "output-" + lang + ".csv" |
| 48 | + if os.path.isdir(db): |
| 49 | + shutil.rmtree(db) |
| 50 | + utils.create_empty_database(lang, ".java", db) |
| 51 | + utils.run_codeql_query(query, db, ql_output) |
| 52 | + |
| 53 | + sources = 0 |
| 54 | + sinks = 0 |
| 55 | + summaries = 0 |
| 56 | + |
| 57 | + with open(ql_output) as csvfile: |
| 58 | + reader = csv.reader(csvfile) |
| 59 | + for row in reader: |
| 60 | + # row: "android.util",1,"remote","source",16 |
| 61 | + if row[3] == "source": |
| 62 | + sources += int(row[4]) |
| 63 | + if row[3] == "sink": |
| 64 | + sinks += int(row[4]) |
| 65 | + if row[3] == "summary": |
| 66 | + summaries += int(row[4]) |
| 67 | + |
| 68 | + os.remove(ql_output) |
| 69 | + |
| 70 | + return (sources, sinks, summaries) |
| 71 | + except: |
| 72 | + print("Unexpected error:", sys.exc_info()[0]) |
| 73 | + raise Exception() |
| 74 | + finally: |
| 75 | + if os.path.isdir(db): |
| 76 | + shutil.rmtree(db) |
| 77 | + |
| 78 | + |
| 79 | +working_dir = "" |
| 80 | +if len(sys.argv) > 1: |
| 81 | + working_dir = sys.argv[1] |
| 82 | + |
| 83 | +configs = [ |
| 84 | + utils.LanguageConfig( |
| 85 | + "java", "Java", ".java", "java/ql/src/meta/frameworks/Coverage.ql") |
| 86 | +] |
| 87 | + |
| 88 | +# todo: change this when we cover multiple languages. We should compute the SHAs |
| 89 | +# only once and not per language |
| 90 | +for config in configs: |
| 91 | + with open("timeseries-" + config.lang + ".csv", 'w', newline='') as csvfile: |
| 92 | + csvwriter = csv.writer(csvfile) |
| 93 | + csvwriter.writerow(["SHA", "Date", "Sources", "Sinks", "Summaries"]) |
| 94 | + |
| 95 | + os.chdir(working_dir) |
| 96 | + |
| 97 | + utils.subprocess_run(["git", "checkout", "main"]) |
| 98 | + |
| 99 | + current_sha = get_str_output(["git", "rev-parse", "HEAD"]) |
| 100 | + current_date = get_date(current_sha) |
| 101 | + |
| 102 | + while True: |
| 103 | + print("Getting stats for " + current_sha) |
| 104 | + utils.subprocess_run(["git", "checkout", current_sha]) |
| 105 | + |
| 106 | + try: |
| 107 | + stats = get_stats(config.lang, config.ql_path) |
| 108 | + |
| 109 | + csvwriter.writerow( |
| 110 | + [current_sha, current_date, stats[0], stats[1], stats[2]]) |
| 111 | + |
| 112 | + print("Collected stats for " + current_sha + |
| 113 | + " at " + current_date.isoformat()) |
| 114 | + except: |
| 115 | + print("Error getting stats for " + |
| 116 | + current_sha + ". Stopping iteration.") |
| 117 | + break |
| 118 | + |
| 119 | + current_sha, current_date = get_previous_sha( |
| 120 | + current_sha, current_date) |
85 | 121 |
|
86 | 122 | utils.subprocess_run(["git", "checkout", "main"]) |
87 | | - |
88 | | - current_sha = get_str_output(["git", "rev-parse", "HEAD"]) |
89 | | - current_date = get_date(current_sha) |
90 | | - |
91 | | - while True: |
92 | | - print("Getting stats for " + current_sha) |
93 | | - utils.subprocess_run(["git", "checkout", current_sha]) |
94 | | - |
95 | | - try: |
96 | | - stats = get_stats() |
97 | | - |
98 | | - csvwriter.writerow( |
99 | | - [current_sha, current_date, stats[0], stats[1], stats[2]]) |
100 | | - |
101 | | - print("Collected stats for " + current_sha + |
102 | | - " at " + current_date.isoformat()) |
103 | | - except: |
104 | | - print("Unexpected error:", sys.exc_info()[0]) |
105 | | - |
106 | | - if os.path.isdir(db): |
107 | | - shutil.rmtree(db) |
108 | | - print("Error getting stats for " + |
109 | | - current_sha + ". Stopping iteration.") |
110 | | - break |
111 | | - |
112 | | - current_sha, current_date = get_previous_sha(current_sha, current_date) |
113 | | - |
114 | | -utils.subprocess_run(["git", "checkout", "main"]) |
|
0 commit comments