Curated OSINT resources for discovering exposed AI infrastructure β dorks, queries, tools, and techniques for LLM, AI agent, and ML pipeline reconnaissance.
`187,000+ Ollama servers exposed` Β· `370,000+ Grok conversations indexed` Β· `AI credential leaks up 81% YoY` Β· `150M+ MCP downloads affected by systemic RCE` Β· `300K+ ChatGPT creds on dark web`
AI OSINT is a curated collection of Google dorks, Shodan queries, GitHub dorks, Censys queries, Sigma detection rules, threat intelligence, and security tools for finding exposed artificial intelligence infrastructure on the internet. It covers LLM endpoints (Ollama, vLLM, LM Studio), AI chatbot conversation leaks (ChatGPT, Grok, Perplexity), vector databases (Qdrant, Weaviate, ChromaDB, Milvus, Pinecone), AI agent gateways (OpenClaw, MCP servers), MLOps platforms (MLflow, Jupyter, Kubeflow), leaked AI API keys (OpenAI, Anthropic, Google Gemini, HuggingFace, Groq, Replicate, Cohere, Mistral, DeepSeek, ElevenLabs), and AI image generation services (Stable Diffusion, ComfyUI). Designed for Red Team operators, penetration testers, bug bounty hunters, and OSINT researchers working in AI/ML security.
It now also covers MCP supply chain attacks (systemic STDIO RCE, marketplace poisoning), AI IDE exploitation (Copilot YOLO mode, Windsurf zero-click), AI model leak incidents (Claude Code source map, DeepSeek ClickHouse exposure), AI-assisted vulnerability discovery (Mythos/Glasswing), and ChatGPT data exfiltration techniques.
Organizations are deploying LLMs, vector databases, and AI agents faster than they can secure them. The result:
- Ollama, vLLM, Gradio β shipped with zero authentication by default
- ChatGPT, Grok β shared conversations indexed by search engines with API keys, passwords, PII
- MCP servers β exposed agent gateways with shell access, file system access, and stored credentials
- Qdrant, ChromaDB, MLflow β no auth out of the box, exposing embeddings, models, and experiments
- AI IDEs (Copilot, Cursor, Windsurf) β prompt injection enables YOLO mode activation, wormable RCE across developer workstations
- MCP supply chain β 9 of 11 MCP marketplaces poisoned in proof-of-concept; 150M+ downloads affected by systemic RCE
- AI code execution sandboxes β DNS side channels bypass outbound network guardrails, enabling silent data exfiltration
This repository gives Red Team operators and OSINT professionals the exact queries to find it all.
Inspired by: 7WaySecurity/cloud_osint
| Section | What You'll Find |
|---|---|
| Google Dorks | Queries for ChatGPT, Grok, HuggingFace, dashboards, config files |
| GitHub Dorks | Leaked API keys for 20+ AI providers, MCP configs, system prompts |
| Shodan Queries | Ollama, vLLM, OpenClaw, Gradio, vector DBs, MLflow, Jupyter |
| Censys Queries | Alternative engine queries for all AI services |
| AI Service Endpoints | URL patterns, default ports, API fingerprinting |
| API Key Patterns | Prefixes, regex, validation for every major AI provider |
| Vector DB Recon | Endpoints to enumerate Qdrant, Weaviate, ChromaDB, Milvus |
| MCP & Agent Exposure | The most critical emerging attack surface in AI security |
| Threat Intelligence | Operation Bizarre Bazaar, Clawdbot crisis, MCP supply chain |
| Tools | AI-specific only β scanners, red team frameworks, key detectors |
| Detection Rules | Sigma rules for monitoring AI infrastructure |
Full collection:
dorks/google/
# Grok (xAI) β 370K+ conversations indexed, NO opt-out for indexing
site:grok.com/share "password"
site:grok.com/share "API key"
site:grok.com/share "secret"
site:grok.com/share "token"
site:grok.com/share "credentials"
site:grok.com/share "keyword"
# ChatGPT β Feature removed Aug 2025, cached results diminishing
# Try on DuckDuckGo which continued indexing after Google stopped
site:chatgpt.com/share "API key"
site:chatgpt.com/share "sk-proj"
site:chatgpt.com/share "password"
site:chatgpt.com/share "AWS_SECRET"
# Perplexity AI
site:perplexity.ai/search "API key"
site:perplexity.ai/search "password"
# Claude (Anthropic) β ~600 convos indexed by Google, 143K+ on Archive.org
# π₯ Original dork by 7WaySecurity
site:claude.ai "public/artifacts"
site:claude.ai/share "API key"
site:claude.ai/share "sk-ant"
site:claude.ai/share "password"
site:claude.ai/share "AWS"
site:claude.ai/share ".env"
site:web.archive.org "claude.ai/share"
# HuggingFace Spaces β keys hardcoded in public Git repos
site:huggingface.co/spaces "sk-proj"
site:huggingface.co/spaces "OPENAI_API_KEY"
site:huggingface.co/spaces "os.environ"
site:huggingface.co/spaces "st.secrets"
intitle:"MLflow" inurl:"/mlflow"
intitle:"Label Studio" inurl:"/projects"
intitle:"Jupyter Notebook" inurl:"/tree" -"Login"
intitle:"Kubeflow" inurl:"/pipeline"
intitle:"Airflow - DAGs"
intitle:"Gradio" inurl:":7860"
intitle:"Streamlit" inurl:":8501"
intitle:"Open WebUI" "ollama"
intitle:"Qdrant Dashboard"
intitle:"ComfyUI"
intitle:"Stable Diffusion"
intitle:"LiteLLM" "proxy"
filetype:env "OPENAI_API_KEY"
filetype:env "ANTHROPIC_API_KEY"
filetype:env "HUGGINGFACE_TOKEN"
filetype:env "GROQ_API_KEY"
filetype:env "PINECONE_API_KEY"
filetype:env "WANDB_API_KEY"
filetype:env "DEEPSEEK_API_KEY"
filetype:env "OPENROUTER_API_KEY"
filetype:yaml "openai" "api_key"
filetype:json "anthropic" "api_key"
# MCP server configs with embedded secrets (systemic RCE β Ox Security, April 2026)
site:github.com "mcpServers" "args" filetype:json
site:github.com "mcp.config" "apiKey"
site:github.com ".cursor" "mcpServers" filetype:json
site:github.com "windsurf" "mcp" "config" filetype:json
# Claude Code leak artifacts
site:github.com "claude-code" "leaked" "source"
# VS Code YOLO mode (CVE-2025-53773)
site:github.com "chat.tools.autoApprove" "true" filetype:json
Full collection:
dorks/github/
β οΈ Syntax note: GitHub migrated to new Code Search. Usepath:*.envinstead of legacyfilename:.env. Queries below use modern syntax where applicable.
# OpenAI (project keys β current format since April 2024)
"sk-proj-" path:*.env
"sk-proj-" path:*.py
"OPENAI_API_KEY" path:*.env NOT "your_key" NOT "example"
# Anthropic
"sk-ant-api03" path:*.env
"ANTHROPIC_API_KEY" path:*.env
# Google AI / Gemini
"AIzaSy" path:*.env "generativelanguage"
"GOOGLE_API_KEY" path:*.env "gemini"
# HuggingFace
"hf_" path:*.env
"HF_TOKEN" path:*.env
# Groq
"gsk_" path:*.env
"GROQ_API_KEY" path:*.env
# Replicate
"r8_" path:*.env
"REPLICATE_API_TOKEN" path:*.env
# Vector DBs & MLOps
"PINECONE_API_KEY" path:*.env
"QDRANT_API_KEY" path:*.env
"WANDB_API_KEY" path:*.env
path:mcp.json "api_key"
path:mcp.json "token"
path:.cursor/mcp.json
"mcpServers" path:*.json "apiKey"
"mcpServers" path:*.json "OPENAI_API_KEY"
"system_prompt" path:*.py "you are"
"SYSTEM_PROMPT" path:*.env
path:prompts.yaml "system"
path:train.jsonl "prompt" "completion"
path:dataset.jsonl "instruction" "output"
# MCP STDIO configs (systemic RCE β 10+ CVEs)
path:*.json "mcpServers" "command" "args"
path:.vscode/settings.json "chat.tools.autoApprove" "true"
# Claude Code attack vectors
"CLAUDE.md" "permission" "allow"
# DeepSeek keys
path:*.env "DEEPSEEK_API_KEY"
"DEEPSEEK_API_KEY" NOT "your_key" NOT "example"
Full collection:
dorks/shodan/
# Ollama β 240,000+ exposed instances worldwide
port:11434 product:"Ollama"
port:11434 http.html:"Ollama"
port:11434 "api/tags"
# vLLM / OpenAI-compatible
port:8000 "openai" "model"
http.title:"FastAPI" port:8000 "/v1/models"
# LM Studio
port:1234 "/v1/models"
# llama.cpp
port:8080 "llama" "completion"
# OpenClaw/Clawdbot β 4,000+ on Shodan, many with zero auth
# Enables RCE via prompt injection, API key theft, reverse shells
http.title:"Clawdbot Control" port:18789
http.title:"OpenClaw" port:18789
port:18789 "api/v1/status"
port:18789 "auth_mode"
http.title:"Gradio" port:7860
http.title:"Streamlit" port:8501
http.title:"Stable Diffusion" port:7860
http.title:"ComfyUI"
# Qdrant β NO auth by default
port:6333 "qdrant"
port:6333 "/collections"
# Weaviate
port:8080 "weaviate"
# Milvus
port:19530 "milvus"
# MLflow β CVE-2026-0545 (CVSS 9.1) RCE, no auth by default
http.title:"MLflow" port:5000
# Jupyter β ~10,000+ on Shodan, targeted by botnets
http.title:"Jupyter Notebook" port:8888 -"Login"
http.title:"JupyterLab" port:8888
# TensorBoard
http.title:"TensorBoard" port:6006
# MCP endpoints (systemic RCE β Ox Security)
http.html:"mcp" "tools" port:3000
http.html:"Model Context Protocol" port:8080
# nginx-ui MCP (CVE-2026-33032 β actively exploited, CVSS 9.8)
http.title:"Nginx UI" port:443
# Flowise (CVE-2026-40933)
http.title:"Flowise" port:3000
# vLLM (LLMjacking target)
http.html:"vLLM" port:8000
# Open WebUI
http.title:"Open WebUI" port:3000
# LiteLLM Proxy
http.title:"LiteLLM" port:4000
# DeepSeek-style ClickHouse exposure
product:"ClickHouse" port:8123
Full collection:
dorks/censys/
# Ollama (Censys found 25%+ on non-default ports)
services.port=11434 AND services.http.response.body:"Ollama"
# Gradio
services.port=7860 AND services.http.response.html_title:"Gradio"
# OpenClaw
services.port=18789 AND services.http.response.body:"Clawdbot"
# Vector DBs
services.port=6333 AND services.http.response.body:"qdrant"
services.port=8080 AND services.http.response.body:"weaviate"
# MLOps
services.port=5000 AND services.http.response.html_title:"MLflow"
services.port=8888 AND services.http.response.html_title:"Jupyter"
### MCP & New AI Infrastructure (v1.2.0)
# MCP endpoints
services.http.response.body:"Model Context Protocol" AND services.port=3000
# Flowise (CVE-2026-40933)
services.http.response.html_title:"Flowise" AND services.port=3000
# vLLM (LLMjacking target)
services.http.response.body:"vLLM" AND services.port=8000
# nginx-ui MCP (CVE-2026-33032 β actively exploited)
services.http.response.html_title:"Nginx UI" AND services.port=443
# ClickHouse exposure (DeepSeek-style)
services.port=8123 AND services.http.response.body:"ClickHouse"
# Open WebUI
services.http.response.html_title:"Open WebUI"
# LiteLLM
services.http.response.html_title:"LiteLLM"
| Provider | API Endpoint | Shared Content |
|---|---|---|
| OpenAI | api.openai.com/v1/* |
chatgpt.com/share/* |
| Anthropic | api.anthropic.com/v1/* |
β |
generativelanguage.googleapis.com/v1beta/* |
β | |
| xAI | api.x.ai/v1/* |
grok.com/share/* |
| Mistral | api.mistral.ai/v1/* |
β |
| Cohere | api.cohere.ai/v1/* |
β |
| DeepSeek | api.deepseek.com/v1/* |
β |
| Groq | api.groq.com/openai/v1/* |
β |
| HuggingFace | api-inference.huggingface.co/models/* |
huggingface.co/spaces/* |
| ElevenLabs | api.elevenlabs.io/v1/* |
β |
| Service | Port | Auth Default | Risk Level |
|---|---|---|---|
| Ollama | 11434 | β None | π΄ Critical |
| vLLM | 8000 | β None | π΄ Critical |
| OpenClaw | 18789 | β None (fixed in rebrand) | π΄ Critical |
| Gradio | 7860 | β None | π΄ Critical |
| Streamlit | 8501 | β None | π‘ High |
| MLflow | 5000 | β None | π΄ Critical |
| Qdrant | 6333/6334 | β None | π΄ Critical |
| Weaviate | 8080 (+ gRPC 50051) | β None | π‘ High |
| ChromaDB | 8000 | β None (binds localhost) | π‘ High |
| Milvus | 19530 (+ mgmt 9091) | β None | π‘ High |
| Jupyter | 8888 | β Token (often disabled) | π‘ High |
| LM Studio | 1234 | β None | π‘ High |
| GPT4All | 4891 | β None | π‘ High |
| Flowise | 3000 | β None | π΄ Critical |
| LiteLLM Proxy | 4000 | β None | π΄ Critical |
| Open WebUI | 3000/8080 | β Auth (default) | π‘ High |
| nginx-ui | 443/9000 | β Auth (bypassable) | π΄ Critical |
| ClickHouse | 8123/9000 | β None | π΄ Critical |
| TGI (HuggingFace) | 8080 | β None | π‘ High |
| Provider | Prefix | Length | Notes |
|---|---|---|---|
| OpenAI | sk-proj- |
~80+ chars | Current format since April 2024 |
| Anthropic | sk-ant-api03- |
~90+ chars | API keys |
| Anthropic OAuth | sk-ant-oat01- |
β | OAuth tokens (Files API, etc.) |
| Google AI | AIzaSy |
39 chars | |
| HuggingFace | hf_ |
~34 chars | Read/write access tokens |
| Replicate | r8_ |
~40 chars | β |
| Groq | gsk_ |
~52 chars | β |
# OpenAI project keys
sk-proj-[A-Za-z0-9_-]{80,}
# Anthropic
sk-ant-api03-[A-Za-z0-9_-]{90,}
# HuggingFace
hf_[A-Za-z0-9]{34}
# Replicate
r8_[A-Za-z0-9]{37}
# Groq
gsk_[A-Za-z0-9]{52}
# Google AI (β οΈ also matches Maps, YouTube, etc.)
AIzaSy[A-Za-z0-9_-]{33}
β οΈ GoogleAIzaSywarning: Google uses the same prefix for ALL Cloud APIs. A Maps API key embedded in public JavaScript can silently gain Gemini API access if the Generative Language API is enabled on the same project. This was documented as a significant issue in early 2026.
When exposed, vector databases leak proprietary embeddings, sensitive documents, and internal knowledge bases used in RAG systems.
| Database | List Collections | Extract Data |
|---|---|---|
| Qdrant | GET /collections |
POST /collections/{name}/points/scroll |
| Weaviate | GET /v1/schema |
GET /v1/objects?class={name} |
| ChromaDB | GET /api/v2/collections |
POST /api/v2/collections/{id}/query |
| Milvus | gRPC ListCollections |
gRPC Search/Query |
β οΈ ChromaDB API update: v1.0.0+ migrated to/api/v2. The legacy/api/v1/collectionsnow returns a deprecation error. Update your recon scripts accordingly.
The Model Context Protocol (MCP) is the most critical emerging attack surface in AI security (2025-2026).
MCP servers connect AI models to shell access, file systems, databases, and APIs. Misconfigurations expose:
- Shell/code execution via prompt injection
- API keys in plaintext in
.envfiles readable by agents - Tool poisoning β hidden instructions in tool metadata
- Supply chain attacks β compromised MCP packages (24,000+ secrets leaked in MCP configs in its first year)
# Shodan
port:18789 "api/v1"
http.title:"Clawdbot" OR http.title:"OpenClaw"
# GitHub
path:mcp.json "apiKey"
path:.cursor/mcp.json
"mcpServers" path:*.json "token"
| Incident | Date | Impact |
|---|---|---|
| OpenClaw Shodan Exposure | Jan 2026 | 4,000+ agent gateways, many with zero auth and RCE |
| Operation Bizarre Bazaar | Dec 2025βJan 2026 | 35,000 attacks on LLM/MCP endpoints; commercial resale |
| Smithery Registry Breach | 2025 | Fly.io token β control of 3,000+ MCP servers |
| mcp-remote CVE-2025-6514 | 2025 | Command injection in 437K+ installs |
| Cursor IDE MCP Trust Issue | 2025 | Persistent RCE via shared repo configs (CVSS 7.2-8.8) |
| MCP Systemic RCE (Ox Security) | Apr 2026 | 150M+ downloads, 200K servers, 10+ CVEs, 9/11 marketplaces poisoned |
| nginx-ui MCPwn (CVE-2026-33032) | Mar 2026 | CVSS 9.8, actively exploited, full Nginx takeover in 2 requests |
| Atlassian MCPwnfluence | Apr 2026 | CVE-2026-27825/27826 β RCE chain from LAN, no auth required |
| MCP TypeScript SDK Data Leak | Apr 2026 | CVE-2026-25536 β cross-client data leak in shared McpServer instances |
| Windsurf Zero-Click RCE | Apr 2026 | CVE-2026-30615 β zero-click prompt injection β local RCE via MCP |
π Full timeline:
threat-intel/THREAT_INTELLIGENCE.md
Full entries:
threat-intel/
- Operation Bizarre Bazaar β First large-scale LLMjacking: 35K attacks, commercial marketplace selling stolen AI access
- OpenClaw/Clawdbot Shodan Crisis β CVE-2026-24061, "Localhost Trust" bypass
- ChatGPT/Grok Conversation Indexing β Google indexed thousands of conversations with credentials
- Claude Conversation Indexing β ~600 conversations indexed by Google (Forbes Sep 2025); 143K+ across all LLMs on Archive.org
- "Claudy Day" Attack Chain β Open redirect + prompt injection + Files API exfiltration in claude.ai (Oasis Security, Mar 2026)
- Claude Code Source Map Leak β 512K lines of source code exposed via npm package v2.1.88 (Mar 2026)
- MCP Supply Chain Timeline β 10+ major breaches in MCP ecosystem
- 175K Ollama Servers Exposed β SentinelOne/Censys study across 130 countries
- AI-Assisted ICS Targeting β 60+ Iranian groups using LLMs for critical infrastructure recon (Feb 2026)
- Claude Mythos Preview / Project Glasswing β Anthropic's unreleased model autonomously discovered thousands of 0-days in major OS/browsers. Limited to ~50 organizations. CVE-2026-4747 (OpenBSD 27-year RCE).
- Claude Code Source Leak β 59.8 MB source map published to npm (Mar 31, 2026). Critical 50-subcommand bypass vulnerability. Fake repos distributing Vidar/GhostSocks malware.
- MCP "Mother of All Supply Chains" β Ox Security found architectural RCE in MCP SDKs: 150M+ downloads, 10+ CVEs, 9/11 marketplaces poisoned. Anthropic declined protocol fix.
- ChatGPT DNS Exfiltration β Check Point discovered silent data leakage via DNS side channel in code execution sandbox. Patched Feb 20, 2026.
- CVE-2025-53773: Copilot Wormable RCE β Prompt injection enables YOLO mode, wormable across repositories. Patched Aug 2025.
- OpenAI Codex CLI Command Injection β Branch injection β lateral movement β codebase access. Patched Feb 2026.
- 300K+ ChatGPT Credentials on Dark Web β IBM X-Force 2026: AI chatbot credentials are emerging infostealer target.
- Ollama: 12,269 More Exposed β LeakIX found additional exposed instances; maintainers continue rejecting auth PRs.
- nginx-ui MCPwn β CVE-2026-33032 actively exploited. Full Nginx takeover in 2 HTTP requests.
- DeepSeek ClickHouse Exposure β Wiz found 1M+ log entries with plaintext chats, API keys, backend details (Jan 2025).
Only AI-specific tools. Full details:
tools/TOOLS.mdFor generic OSINT tools (Shodan, Censys, etc.) see cloud_osint.
| Tool | What It Does | By |
|---|---|---|
| Garak | LLM vulnerability scanner β "nmap for LLMs" | NVIDIA |
| PyRIT | AI red teaming framework β Crescendo, jailbreaking, multi-turn | Microsoft |
| promptfoo | LLM pentesting CLI β 133+ attack plugins | OpenAI |
| DeepTeam | 50+ vulns, 20+ attacks, OWASP/MITRE mapping | Confident AI |
| API Radar | Real-time leaked AI API key monitoring on GitHub | Independent |
| KeyLeak Detector | Web scanner for 200+ patterns incl. 15+ AI providers | Independent |
| promptmap | ChatGPT dorks & prompt injection testing | Utku Εen |
| Vulnerable MCP | MCP vulnerability database with CVEs and PoCs | Community |
| MCPSafetyScanner | MCP server security auditing tool | Academic (Leidos) |
| Cisco AI Supply Chain Scanners | Scanners for MCP, A2A, pickle files, agentic skill files | Cisco Talos |
| DorkEye | Automated Google Dorking with multi-agent analysis pipeline | Open Source |
Full Sigma rules:
detection-rules/SIGMA_RULES.md
12 Sigma detection rules covering:
- External access to Ollama (port 11434)
- AI API keys appearing in application logs
- Unauthorized LLM API access without Bearer tokens
- Exposed MCP servers with
auth_mode: none - Vector database unauthorized enumeration
- LLMjacking β anomalous inference spikes (>500 requests/hour)
- AI agent RCE via prompt injection (suspicious tool calls to bash/python_repl)
- MCP STDIO arbitrary command execution (CVE-2026-30615/30624/30616/40933)
- VS Code Copilot YOLO mode activation (CVE-2025-53773)
- nginx-ui MCP endpoint unauthenticated access (CVE-2026-33032, actively exploited)
- DNS-based data exfiltration from AI code execution sandboxes
- Claude Code 50+ subcommand pipeline deny rule bypass
| Source | Finding | Link |
|---|---|---|
| SentinelOne + Censys | 175,108 exposed Ollama hosts across 130 countries | SecurityWeek |
| Cisco Talos | 1,100+ Ollama servers, 20% serving models without auth | Cisco Blog |
| Pillar Security | Operation Bizarre Bazaar: 35,000 LLMjacking attacks | Pillar Security |
| GitGuardian | AI credential leaks surged 81% YoY; 29M secrets on GitHub | Report |
| AuthZed | Complete MCP security breaches timeline | Blog |
| Pangea | Sensitive data in indexed ChatGPT histories | Blog |
| Trail of Bits | 8 high-severity vulns in Gradio 5 audit | Blog |
| Oasis Security | "Claudy Day" β 3 vulns chained for data exfiltration from Claude | Blog |
| Forbes | ~600 Claude conversations indexed by Google | Article |
| Obsidian Security | 143K+ LLM chats (incl. Claude) on Archive.org | Blog |
| TechRadar | Claude Code 512K source lines leaked via npm | Article |
| Penligent | Claude Code source map leak analysis | Blog |
| Ox Security | 150M+ MCP downloads affected by systemic RCE, 10+ CVEs, 9/11 marketplaces poisoned | Blog |
| Anthropic Red Team | Claude Mythos: thousands of 0-days, CVE-2026-4747, sandbox escape | red.anthropic.com |
| IBM X-Force | 300K+ ChatGPT creds on dark web, supply chain attacks 4x in 5 years | Report |
| Check Point | ChatGPT DNS exfiltration β silent data leakage via side channel | Blog |
| Zscaler ThreatLabz | Claude Code source leak + Vidar/GhostSocks malware lure | Blog |
| Cisco AI | State of AI Security 2026 β MCP, A2A, agentic AI scanners | Report |
| LeakIX | 12,269 additional exposed Ollama instances, auth PR rejection critique | Blog |
| CSA | Time-to-exploit now under 20 hours (Mythos briefing) | HelpNetSecurity |
| Wiz | Claude Mythos practical guidance β "Y2K moment" for cybersecurity | Blog |
| Wiz | DeepSeek ClickHouse exposure β 1M+ log entries with plaintext chats | Blog |
- OWASP Top 10 for LLM Applications 2025
- OWASP Top 10 for Agentic Applications 2026
- MITRE ATLAS
- MCP Security Best Practices
- vulnerablemcp.info
The resources in this repository are for authorized security testing, education, and legitimate research only.
- Obtain proper authorization before testing infrastructure you do not own
- Follow applicable laws (CFAA, GDPR, local equivalents)
- Report vulnerabilities responsibly through appropriate disclosure channels
- Never exploit discovered misconfigurations for unauthorized access
The authors assume no liability for misuse.
PRs welcome! Please verify dorks/queries work before submitting. See CONTRIBUTING.md.
Code: MIT Β· Data & Documentation: CC BY-SA 4.0
Maintained by 7WaySecurity Β· Last updated April 2026
