Skip to content

spidybroz/PC1

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

56 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

SpidyCrawler - Synthetic Web Traffic Agent v1.2

Python Platform License Status

β€œSimulate real users, understand real behavior, optimize real performance.”

A sophisticated Python-based web traffic bot that simulates human-like browsing sessions to visit websites with natural interactions.

πŸ“ Project Structure

SpidyCrawler/
β”œβ”€β”€ runBot.bat                                  # Perfect launcher
β”œβ”€β”€ Source/
β”‚   β”œβ”€β”€ __pycache__/                            # Cache
β”‚   β”œβ”€β”€ SC_BOT.py                              # Updated for multi-bot
β”‚   └── bot_manager.py                          # Unlimited scaling
β”œβ”€β”€ Customize/
β”‚   β”œβ”€β”€ urls.txt
β”‚   β”œβ”€β”€ spend_time.txt
β”‚   └── bot_count.txt                           # Your bot count (49, 100, 1000, etc.)
└── Logs/
    β”œβ”€β”€ bot_1/                                  # Individual bot logs
    β”œβ”€β”€ bot_2/
    └── ... (as many as bot_count.txt)

πŸš€ Quick Start – One File Does Everything

  1. Double-click runBot.bat (or run python run.py from the project folder).
  2. The first run will automatically:
    • Create Customize/ and default config files (urls.txt, spend_time.txt, bot_count.txt) if missing
    • Install dependencies (pip install -r source/requirements.txt)
    • Fetch free proxies if proxies.txt is empty
    • Start the bot(s)

No manual setup. Edit Customize/urls.txt (and other files) after the first run if you want to change targets or bot count.

Run on another PC (auto setup, auto config, auto download)

  1. Copy the whole SpidyCrawler folder to the other PC (or clone the repo).
  2. On that PC install Python 3.8+ (python.org) and Chrome (for headless browsing). Add Python to PATH when installing.
  3. Double-click runBot.bat (or run python run.py).
  4. First run will automatically: upgrade pip, install dependencies (with one retry), pin blinker for selenium-wire, create default config if missing, fetch free proxies if proxies.txt is empty, then start the bot(s). No manual pip or config steps needed.

βš™οΈ Configuration

I'll update the bot count section of your documentation. Here's the revised configuration section:

βš™οΈ Configuration

1. BOT Count (Customize/bot_count.txt)

System Requirements Guide:

Specify the number of concurrent bots to run
10    # Safe test          (~2GB RAM, 4-core CPU)
50    # Balanced scale     (~10GB RAM, 6-core CPU) 
100   # Medium scale       (~20GB RAM, 8-core CPU)  
500   # Large scale        (~100GB RAM, 16-core CPU)
1000  # Extreme scale      (Server-grade hardware)
5000  # INSANE scale! πŸš€

Performance Optimizations:

  • Smart Concurrency: Higher concurrency limits for more bots
  • Reduced Console Spam: Only shows first 50 bots in console
  • Progress Tracking: For 50+ bots, shows progress every 10 completions
  • Memory Warnings: Automatic RAM usage estimates
  • Safety Confirmation: Asks for confirmation for 1000+ bots

2. Visit Duration (Customize/spend_time.txt)

Specify time in seconds (60 seconds to 24 hours):

600 seconds

This means 10 minutes (600 seconds) per URL.

3. URLs Configuration (Customize/urls.txt)

Add your target URLs, one per line:

https://www.example.com/
https://www.anotherexample.com/
https://www.yoursite.com/

4. Proxies – Different IP/Location Worldwide (Optional, 100% Free)

Traffic can appear from different IPs/locations. Each bot gets a proxy (round-robin). Production-ready free flow:

  • Auto (recommended): Leave Customize/proxies.txt missing or empty. On first run, the bot fetches free proxies from public lists and saves them. No signup, no payment.
  • Manual refresh: Run python source/proxy_fetcher.py to fetch/refresh free proxies. Use --no-validate for a faster list (more entries, some may be dead).
  • Custom: Create Customize/proxies.txt with one proxy per line (http://ip:port or http://user:pass@host:port, socks5://…). Auth is supported via selenium-wire.
  • Cycle: Fewer proxies than bots β†’ proxies repeat (bot 1 β†’ proxy 1, bot 2 β†’ proxy 2, …).
  • No proxy: If fetch fails or you remove the file, bots run without proxy.

🎯 Features

πŸ€– Human-Like Behavior

  • Natural browsing patterns with random delays
  • Realistic mouse movements and scrolling
  • Smart element clicking on buttons and links
  • Variable session durations based on total time

πŸ”§ Advanced Capabilities

  • Redirect handling - Maintains exact URLs
  • Anti-detection - Random user agents and browser fingerprints
  • Verification detection - Identifies CAPTCHAs and challenges
  • Comprehensive logging - Detailed activity tracking

⏱️ Smart Time Management

  • Short visits (1-5 min): Quick interactions
  • Medium visits (5-30 min): Balanced activities
  • Long visits (30+ min): Extended browsing sessions

πŸ“Š Logging & Monitoring

The bot creates detailed logs in the Logs/bot1.... folder:

  • SC_BOT.log - Real-time activity log
  • session_progress.json - Live progress tracking
  • session_final.json - Complete session summary
  • verification_challenges.log - Security challenge records

πŸ›‘οΈ Anti-Detection Features

  • Random User Agents - Different browsers and devices
  • Browser Fingerprint Spoofing - Avoids automation detection
  • Natural Timing - Human-like delays between actions
  • Realistic Interactions - Mouse movements, scrolling, clicking

⚠️ Important Notes

βœ… Recommended Usage

  • Legitimate testing of your own websites
  • SEO monitoring and analytics verification
  • Load testing and user behavior simulation
  • Educational purposes for web analytics

❌ Prohibited Usage

  • Illegal activities or harassment
  • Competitor manipulation or malicious intent
  • Spam generation or abusive behavior
  • Terms of service violations

⚑ Performance Tips

  • Start small with 1-2 URLs and short durations
  • Monitor logs for any verification challenges
  • Use responsibly to avoid overwhelming servers
  • Respect robots.txt and website policies

πŸ”§ Technical Requirements

  • Windows 10/11 (batch script optimized for Windows)
  • Python 3.11+ (automatically installed if missing)
  • Chrome Browser (required for Selenium WebDriver)
  • Internet Connection (for downloads and browsing)

πŸ“‹ Dependencies

The bot automatically installs:

  • selenium - Web browser automation
  • fake-useragent - Random user agent generation
  • urllib3 - URL handling utilities
  • selenium-wire - Proxy support (including auth) for different IP/location

πŸ†˜ Troubleshooting

Common Issues

  1. Python not found after installation

    • Solution: Restart your computer and run runBot.bat again
  2. Chrome driver issues

    • Solution: Ensure Chrome is installed and updated
  3. Verification challenges detected

    • Solution: The bot will log these and may require manual intervention
  4. URL redirects happening

    • Solution: The bot automatically detects and attempts to fix redirects

Log Files Location

Check SC_BOT\Logs\ for detailed error information and session reports.

πŸ“„ License

This project is for educational and legitimate testing purposes only. Users are responsible for complying with all applicable laws and website terms of service.

🀝 Contributing

For improvements and bug reports, please ensure all contributions adhere to ethical usage guidelines.


Remember: Use this tool responsibly and only on websites you own or have explicit permission to test. Always respect website terms of service and robots.txt directives.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •