210 lines
6.1 KiB
Python
Executable file
210 lines
6.1 KiB
Python
Executable file
#!/usr/bin/env python3
|
|
"""
|
|
GitHub PR Analyzer for zed-industries/zed repository
|
|
Downloads all PRs and groups them by first assignee with status, open date, and last updated date.
|
|
"""
|
|
|
|
import urllib.request
|
|
import urllib.parse
|
|
import urllib.error
|
|
import json
|
|
from datetime import datetime
|
|
from collections import defaultdict
|
|
import sys
|
|
import os
|
|
|
|
# GitHub API configuration
|
|
GITHUB_API_BASE = "https://api.github.com"
|
|
REPO_OWNER = "zed-industries"
|
|
REPO_NAME = "zed"
|
|
GITHUB_TOKEN = os.getenv("GITHUB_TOKEN")
|
|
|
|
def make_github_request(url, params=None):
|
|
"""Make a request to GitHub API with proper headers and pagination support."""
|
|
if params:
|
|
url_parts = list(urllib.parse.urlparse(url))
|
|
query = dict(urllib.parse.parse_qsl(url_parts[4]))
|
|
query.update(params)
|
|
url_parts[4] = urllib.parse.urlencode(query)
|
|
url = urllib.parse.urlunparse(url_parts)
|
|
|
|
req = urllib.request.Request(url)
|
|
req.add_header("Accept", "application/vnd.github.v3+json")
|
|
req.add_header("User-Agent", "GitHub-PR-Analyzer")
|
|
|
|
if GITHUB_TOKEN:
|
|
req.add_header("Authorization", f"token {GITHUB_TOKEN}")
|
|
|
|
try:
|
|
response = urllib.request.urlopen(req)
|
|
return response
|
|
except urllib.error.URLError as e:
|
|
print(f"Error making request to {url}: {e}")
|
|
return None
|
|
except urllib.error.HTTPError as e:
|
|
print(f"HTTP error {e.code} for {url}: {e.reason}")
|
|
return None
|
|
|
|
def fetch_all_prs():
|
|
"""Fetch all PRs from the repository using pagination."""
|
|
prs = []
|
|
page = 1
|
|
per_page = 100
|
|
|
|
print("Fetching PRs from GitHub API...")
|
|
|
|
while True:
|
|
url = f"{GITHUB_API_BASE}/repos/{REPO_OWNER}/{REPO_NAME}/pulls"
|
|
params = {
|
|
"state": "open",
|
|
"sort": "updated",
|
|
"direction": "desc",
|
|
"per_page": per_page,
|
|
"page": page
|
|
}
|
|
|
|
response = make_github_request(url, params)
|
|
if not response:
|
|
break
|
|
|
|
try:
|
|
data = response.read().decode('utf-8')
|
|
page_prs = json.loads(data)
|
|
except (json.JSONDecodeError, UnicodeDecodeError) as e:
|
|
print(f"Error parsing response: {e}")
|
|
break
|
|
|
|
if not page_prs:
|
|
break
|
|
|
|
prs.extend(page_prs)
|
|
print(f"Fetched page {page}: {len(page_prs)} PRs (Total: {len(prs)})")
|
|
|
|
# Check if we have more pages
|
|
link_header = response.getheader('Link', '')
|
|
if 'rel="next"' not in link_header:
|
|
break
|
|
|
|
page += 1
|
|
|
|
print(f"Total PRs fetched: {len(prs)}")
|
|
return prs
|
|
|
|
def format_date_as_days_ago(date_string):
|
|
"""Format ISO date string as 'X days ago'."""
|
|
if not date_string:
|
|
return "N/A days ago"
|
|
|
|
try:
|
|
dt = datetime.fromisoformat(date_string.replace('Z', '+00:00'))
|
|
now = datetime.now(dt.tzinfo)
|
|
days_diff = (now - dt).days
|
|
|
|
if days_diff == 0:
|
|
return "today"
|
|
elif days_diff == 1:
|
|
return "1 day ago"
|
|
else:
|
|
return f"{days_diff} days ago"
|
|
except:
|
|
return "N/A days ago"
|
|
|
|
def get_first_assignee(pr):
|
|
"""Get the first assignee from a PR, or return 'Unassigned' if none."""
|
|
assignees = pr.get('assignees', [])
|
|
if assignees:
|
|
return assignees[0].get('login', 'Unknown')
|
|
return 'Unassigned'
|
|
|
|
def get_pr_status(pr):
|
|
"""Determine if PR is draft or ready for review."""
|
|
if pr.get('draft', False):
|
|
return "Draft"
|
|
return "Ready"
|
|
|
|
def analyze_prs(prs):
|
|
"""Group PRs by first assignee and organize the data."""
|
|
grouped_prs = defaultdict(list)
|
|
|
|
for pr in prs:
|
|
assignee = get_first_assignee(pr)
|
|
|
|
pr_info = {
|
|
'number': pr['number'],
|
|
'title': pr['title'],
|
|
'status': get_pr_status(pr),
|
|
'state': pr['state'],
|
|
'created_at': format_date_as_days_ago(pr['created_at']),
|
|
'updated_at': format_date_as_days_ago(pr['updated_at']),
|
|
'updated_at_raw': pr['updated_at'],
|
|
'url': pr['html_url'],
|
|
'author': pr['user']['login']
|
|
}
|
|
|
|
grouped_prs[assignee].append(pr_info)
|
|
|
|
# Sort PRs within each group by update date (newest first)
|
|
for assignee in grouped_prs:
|
|
grouped_prs[assignee].sort(key=lambda x: x['updated_at_raw'], reverse=True)
|
|
|
|
return dict(grouped_prs)
|
|
|
|
def print_pr_report(grouped_prs):
|
|
"""Print formatted report of PRs grouped by assignee."""
|
|
print(f"OPEN PR REPORT FOR {REPO_OWNER}/{REPO_NAME}")
|
|
print()
|
|
|
|
# Sort assignees alphabetically, but put 'Unassigned' last
|
|
assignees = sorted(grouped_prs.keys())
|
|
if 'Unassigned' in assignees:
|
|
assignees.remove('Unassigned')
|
|
assignees.append('Unassigned')
|
|
|
|
total_prs = sum(len(prs) for prs in grouped_prs.values())
|
|
print(f"Total Open PRs: {total_prs}")
|
|
print()
|
|
|
|
for assignee in assignees:
|
|
prs = grouped_prs[assignee]
|
|
assignee_display = f"@{assignee}" if assignee != 'Unassigned' else assignee
|
|
print(f"assigned to {assignee_display} ({len(prs)} PRs):")
|
|
|
|
for pr in prs:
|
|
print(f"- {pr['author']}: [{pr['title']}]({pr['url']}) opened:{pr['created_at']} updated:{pr['updated_at']}")
|
|
|
|
print()
|
|
|
|
def save_json_report(grouped_prs, filename="pr_report.json"):
|
|
"""Save the PR data to a JSON file."""
|
|
try:
|
|
with open(filename, 'w') as f:
|
|
json.dump(grouped_prs, f, indent=2)
|
|
print(f"📄 Report saved to {filename}")
|
|
except Exception as e:
|
|
print(f"Error saving JSON report: {e}")
|
|
|
|
def main():
|
|
"""Main function to orchestrate the PR analysis."""
|
|
print("GitHub PR Analyzer")
|
|
print("==================")
|
|
|
|
if not GITHUB_TOKEN:
|
|
print("⚠️ Warning: GITHUB_TOKEN not set. You may hit rate limits.")
|
|
print(" Set GITHUB_TOKEN environment variable for authenticated requests.")
|
|
print()
|
|
|
|
# Fetch all PRs
|
|
prs = fetch_all_prs()
|
|
|
|
if not prs:
|
|
print("❌ Failed to fetch PRs. Please check your connection and try again.")
|
|
sys.exit(1)
|
|
|
|
# Analyze and group PRs
|
|
grouped_prs = analyze_prs(prs)
|
|
|
|
# Print report
|
|
print_pr_report(grouped_prs)
|
|
|
|
if __name__ == "__main__":
|
|
main()
|