Find potential customers from GitHub stargazer activity. Monitor any repositories for
new stars, scrape historic stargazers on first run, and use AI to qualify leads automatically.
What this does
Historic Scraping: On first run, scrapes all existing stargazers from configured repositories
Ongoing Monitoring: A cron job monitors for new stars on your configured repositories
AI Lead Qualification: An AI agent researches each stargazer and scores them as a lead
Dashboard & Alerts: Qualified leads appear in your dashboard with periodic digest emails
Sends daily emails (1pm UTC) with new qualified leads
Edit RECIPIENTS array to configure who receives them
Customization
Repositories: Edit the GITHUB_REPOS array in github.cron.ts to add/remove repositories
Lead Criteria: Edit PROMPT.txt to define your ideal customer profile
Monitoring Frequency: Adjust the cron schedule in github.cron.ts (default: hourly)
Email Schedule: Adjust the email digest schedule in digest.ts (default: daily at 1pm UTC)
Repository Configuration
Add repositories to monitor in the GITHUB_REPOS array:
exportconstGITHUB_REPOS = [
"your-org/your-repo",
"facebook/react",
"microsoft/vscode",
"octocat/Hello-World", // Good for testing// Add more repositories here
];
Performance Notes
Gradual Scraping: Large repositories are processed over multiple cron runs to avoid timeouts
Rate Limits: The system includes delays and batching to handle GitHub API limits
Incremental Updates: After historic scraping, only new stars are processed
Error Handling: Individual lead processing failures won't stop the entire job
Resume Capability: System automatically resumes scraping after interruptions
Important Limits & Scraping Strategy
Small Repos (≤1,000 stars): Complete historic scraping in one run
Medium Repos (1,001-5,000 stars): Scrape 500 stars per cron run
Large Repos (5,001-10,000 stars): Scrape 200 stars per cron run (gradual)
Very Large Repos (>10,000 stars): Skip historic scraping, monitor new stars only
Batch Processing: Leads are processed in batches of 5 to ensure reliability
Progress Tracking: System resumes scraping where it left off across multiple runs