|
79 | 79 | GITHUB_TOKEN = os.getenv("GITHUB_TOKEN")
|
80 | 80 | ORG_NAME = None # Will be set by argparse
|
81 | 81 |
|
| 82 | +# GitHub API headers |
| 83 | +HEADERS = { |
| 84 | + "Authorization": f"Bearer {GITHUB_TOKEN}", |
| 85 | + "Accept": "application/vnd.github+json", |
| 86 | +} |
| 87 | + |
82 | 88 | # List of repositories to exclude in the format 'org/repo'
|
83 | 89 | EXCLUDED_REPOS = [
|
84 | 90 | "pytorch/pytorch",
|
@@ -150,7 +156,7 @@ def get_repos(org: str) -> List[str]:
|
150 | 156 | while True:
|
151 | 157 | url = f"{BASE_URL}/orgs/{org}/repos?per_page=100&page={page}"
|
152 | 158 | logging.debug(f"[get_repos] Requesting URL: {url}")
|
153 |
| - data = make_cached_request(url) |
| 159 | + data = make_cached_request(url, HEADERS) |
154 | 160 | if data is None:
|
155 | 161 | logging.error(f"[get_repos] Failed to fetch page {page} for org: {org}")
|
156 | 162 | break
|
@@ -186,7 +192,7 @@ def get_workflow_runs(org: str, repo: str) -> List[Dict]:
|
186 | 192 | while True:
|
187 | 193 | url = f"{BASE_URL}/repos/{org}/{repo}/actions/runs?per_page=100&page={page}&created=>={WORKFLOW_RUN_LOOKBACK}"
|
188 | 194 | logging.debug(f"[get_workflow_runs] Requesting URL: {url}")
|
189 |
| - response_data = make_cached_request(url) |
| 195 | + response_data = make_cached_request(url, HEADERS) |
190 | 196 | if response_data is None:
|
191 | 197 | logging.error(
|
192 | 198 | f"[get_workflow_runs] Failed to fetch page {page} for repo: {repo}"
|
@@ -271,7 +277,7 @@ def get_jobs_for_run(
|
271 | 277 | )
|
272 | 278 | url = f"{BASE_URL}/repos/{org}/{repo}/actions/runs/{run_id}/jobs"
|
273 | 279 | logging.debug(f"[get_jobs_for_run] Requesting URL: {url}")
|
274 |
| - response_data = make_cached_request(url) |
| 280 | + response_data = make_cached_request(url, HEADERS) |
275 | 281 | if response_data is None:
|
276 | 282 | logging.error(
|
277 | 283 | f"[get_jobs_for_run] Failed to fetch jobs for run {run_id} in repo: {repo}"
|
|
0 commit comments