Google Dorks List and Updated Database for Web Server Detection in 2021

11 min read
Last updated: Apr 8, 2021

google-dork-list

How to Use Google Dorks?

To use a Google Dork, you simply type in a Dork into the search box on Google and press “Enter”. Here are some of the best Google Dork queries that you can use to search for information on Google.

Google Dork Queries Examples:

  • site:.edu “phone number”– This Dork searches for websites on .edu domains that contain the words “phone number”. student “phone number” – This Dork searches for websites on .edu domains that contain the words “student” and “phone number”.
  • inurl:edu “login” – This Dork searches for websites on .edu domains that contain the words “login”. This Dork searches for school websites that contain student login information.
  • “powered by vbulletin” site:.edu – This Dork searches for websites on .edu domains that contain the words “powered by vbulletin”. This Dork searches for school websites that are running on the vbulletin forum software.
  • “powered by vbulletin” site:.gov – This Dork searches for websites on .gov domains that contain the words “powered by vbulletin”. This Dork searches for governmental websites that are running on the vbulletin forum software.
  • “powered by vbulletin” site:.mil – This Dork searches for websites on .mil domains that contain the words “powered by vbulletin”. This Dork searches for military websites that are running on the vbulletin forum software.
  • “powered by vbulletin” inurl:.edu – This Dork searches for websites on .edu domains that contain the words “powered by vbulletin”. This Dork searches for school websites that are running on the vbulletin forum software.
  • “powered by vbulletin” inurl:.mil – This Dork searches for websites on .mil domains that contain the words “powered by vbulletin”. This Dork searches for military websites that are running on the vbulletin forum software.
  • inurl:.com “powered by vbulletin” – This Dork searches for websites on .com domains that contain the words “powered by vbulletin”. This Dork searches for websites that are running on the vbulletin forum software.
  • inurl:.edu “register forum” – This Dork searches for websites on .edu domains that contain the words “register forum”. This Dork searches for school websites that allow you to register for a forum.
  • inurl:.gov “register forum” – This Dork searches for websites on .gov domains that contain the words “register forum”. This Dork searches for governmental websites that allow you to register for a forum.

Scraper API provides a proxy service designed for web scraping. With over 20 million residential IPs across 12 countries, as well as software that can handle JavaScript rendering and solving CAPTCHAs, you can quickly complete large scraping jobs without ever having to worry about being blocked by any servers.

Google Dorks Updated Database for Web Server Detection:

>inurl:/uploads/affwp-debug.log
>inurl:?XDEBUG_SESSION_START=phpstorm
>inurl:/config/device/wcd
>inurl:"/phpmyadmin/user_password.php
>inurl:/xprober ext:php
>intext:"SonarQube" + "by SonarSource SA." + "LGPL v3"
>intext:"Healthy" + "Product model" + " Client IP" + "Ethernet"
>inurl:/phpPgAdmin/browser.php
>ext:php | intitle:phpinfo "published by the PHP Group"
>allintext:"Index Of" "sftp-config.json"
>inurl:_vti_bin/Authentication.asmx
>"Powered by 123LogAnalyzer"
>intitle:Snoop Servlet
>allintitle:"Pi-hole Admin Console"
>intitle:"Lists Web Service"
>intitle:"Monsta ftp" intext:"Lock session to IP"
>intitle:"Microsoft Internet Information Services 8" -IIS
>intext:"index of /" "Index of" access_log
>inurl:"id=*" & intext:"warning mysql_fetch_array()"
>"index of /private" -site:net -site:com -site:org
>inurl:":8088/cluster/apps"
>intitle:"index of" "docker.yml"
>intitle:"index of" "debug.log" OR "debug-log"
>intext:"This is the default welcome page used to test the correct operation of the Apache2 server"
>"Powered by phpBB" inurl:"index.php?s" OR inurl:"index.php?style"
>intitle:"index of" "powered by apache " "port 80"
>intitle:"Web Server's Default Page" intext:"hosting using Plesk" -www
>site:ftp.*.com "Web File Manager"
>intitle:"Welcome to JBoss"
>intitle:"Welcome to nginx!" intext:"Welcome to nginx on Debian!" intext:"Thank you for"
>intitle:"index of" "Served by Sun-ONE"
>-pub -pool intitle:"index of" "Served by" "Web Server"
>intitle:"index of" "server at"
>site:*/server-status intext:"Apache server status for"
>inurl:/iisstart.htm intitle:"IIS7"
>inurl:/phpmyadmin/changelog.php -github -gitlab
>inurl:"WebPortal?bankid"
>intitle:"IIS Windows Server" -inurl:"IIS Windows Server"
>intitle:"Apache2 Ubuntu Default Page: It works"
>inurl:/server-status + "Server MPM:"
>inurl:phpinfo.php intext:build 2600
>intext:"Brought to you by eVetSites"
>inurl:OrganizationChart.cc
>intext:"Powered by GetSimple" -site:get-simple.info
>inurl:jsmol.php
>intitle:"WAMPSERVER homepage" "Server Configuration" "Apache Version"
>"Powered by BOINC"
>"Proudly created with Wix.com"
>"I have been invoked by servletToJSP"
>intext:" - 2019 Cott Systems, Inc."
>intitle:Test Page for the Nginx HTTP Server on Fedora
>"Powered by vShare"
>intext:"Powered by Abyss Web Server"
>intitle: "Welcome to nginx!" + "Thank you for using nginx."
>intitle:livezilla "Server Time"
>intext:"Powered by phpSQLiteCMS" | intitle:"phpSQLiteCMS - A simple & lightweight CMS"
>intitle:"Sucuri WebSite Firewall - Access Denied"
>"This server is operated by OpenX."
>intext:"Powered by (Quantum | Quantum CMS | CMS)
>intext:"Powered by Typesetter"
>Coldbox | contentbox | commandbox "Powered by ContentBox"
>intext:"Powered by Nesta"
>intext:"Powered by Sentora" -github.com
>inurl:/_hcms/
>inurl:tests/mocks intext:autoloader
>intitle:HTTP Server Test Page powered by CentOS
>intitle:"apache tomcat/" "Apache Tomcat examples"
>inurl:composer.json codeigniter -site:github.com
>"RDServer Product information" | inurl:"/rdagent.jsp"
>intitle:"apache tomcat/" + "Find additional important configuration information in:"
>intitle:"Apache2 Debian Default Page: It works"
>inurl:rvsindex.php & /rvsindex.php?/user/login
>intitle:"miniProxy"
>intitle:"Axis Happiness Page" "Examining webapp configuration"
>"var miner = new CoinHive" intext:document.domain
>inurl:"server-status" "Server Version: Apache/" "Server Built: " "Server uptime:" "Total accesses" "CPU Usage:"
>"Application Blocked!" "Google bot"
>inurl:readme.md intext:"Laravel"
>inurl:phpmyadmin/themes intext:"pmahomme"
>intext:"Welcome to CodeIgniter!"
>inurl:readme.rst intext:"CodeIgniter"
>intitle:"Index of" "Apache/2.4.7 (Ubuntu) Server"
>intitle:"Index of /" "Proudly Served by Surftown at"
>inurl:/pub/ inurl:_ri_
>inurl:/_catalogs
>ext:svc inurl:wsdl
>inurl:user_guide intext:"CodeIgniter User Guide"
>"PHP Credits" "Configuration" "PHP Core" ext:php inurl:info
>inurl:/php/info.php
>intitle:"HFS" "Server Uptime" "Server time"
>inurl:phpsysinfo/index.php?disp=dynamic
>intitle:"Apache Status" | intext:"Apache Server Status"
>intext:Apache/2.2.29 (Unix) mod_ssl/2.2.29 | intitle:"Index of /"
>inurl:"/web-console/" intitle:"Administration Console"
>filetype:asmx inurl:(_vti_bin|api|webservice)
>intitle:"BadBlue: the file-sharing web server anyone can use"
>intitle:"Apache Status" "Apache Server Status for"
>intext:"Target Multicast Group" "beacon"
>inurl:wl.exe inurl:?SS1= intext:"Operating system:" -edu -gov -mil
>inurl:nnls_brand.html OR inurl:nnls_nav.html
(intitle:"502 Proxy Error")|(intitle:"503 Proxy Error") "The proxy server could not handle the request" -topic -mail -4suite -list -site:geocrawler.co
intitle:"Welcome to 602LAN SUITE *"
intitle:"Welcome To Your WebSTAR Home Page"
intitle:"Document title goes here" intitle:"used by web search tools" " example of a simple Home Page"
intitle:"Welcome to the Advanced Extranet Server, ADVX!"
intitle:"Welcome to Windows Small Business Server 2003"
yaws.*.server.at
intitle:"Index of *" mode links bytes last-changed name
intitle:"IPC@CHIP Infopage"
intitle:"Test Page for the Apache HTTP Server on Fedora Core" intext:"Fedora Core Test Page"
Powered.by.RaidenHTTPD intitle:index.of
(inurl:81-cobalt | inurl:cgi-bin/.cobalt)
intitle:"welcome to mono xsp"
inurl:oraweb -site:oraweb.org
"Netware * Home" inurl:nav.html
XAMPP "inurl:xampp/index"
inurl:2506/jana-admin
allintext:"Powered by LionMax Software" "WWW File Share"
intitle:"Resin Default Home Page"
intitle:"Welcome To Xitami" -site:xitami.com

A Google Dork is a search query that looks for specific information on Google’s search engine. Google Dorks are developed and published by hackers and are often used in “Google Hacking”.

Google Dorks are extremely powerful. They allow you to search for a wide variety of information on the internet and can be used to find information that you didn’t even know existed.

Because of the power of Google Dorks, they are often used by hackers to find information about their victims or to find information that can be used to exploit vulnerabilities in websites and web applications.

Google Search Engine is designed to crawl anything over the internet and this helps us to find images, text, videos, news and plethora of information sources. With it’s tremendous capability to crawl, it indexes data along the way, which also includes sensitive information like email addresses, login credentials, sensitive files, website vulnerabilities, and even financial information. In many cases, We as a user won’t be even aware of it.

Google Dork is a search query that we give to Google to look for more granular information and retrieve relevant information quickly. For example, try to search for your name and verify results with a search query [inurl:your-name]. Analyse the difference. You just have told google to go for a deeper search and it did that beautifully.

Search Operators :

  • cache:

    • [cache:www.google.com web] will show the cached content with the word “web” highlighted. This functionality is also accessible by clicking on the “Cached” link on Google’s main results page. The query [cache:] will show the version of the web page that Google has in its cache. For instance, [cache:www.google.com] will show Google’s cache of the Google homepage.
  • link:

    • [link:www.google.com] will list webpages that have links pointing to the Google homepage.
  • related:

    • [related:www.google.com] will list web pages that are similar to the Google homepage.
  • info:

    • [info:www.google.com] will show information about the Google homepage.
  • define:

    • The query [define:] will provide a definition of the words you enter after it, gathered from various online sources. The definition will be for the entire phrase entered (i.e., it will include all the words in the exact order you typed them). Eg: [define:google]
  • stocks:

    • If you begin a query with the [stocks:] operator, Google will treat the rest of the query terms as stock ticker symbols, and will link to a page showing stock information for those symbols. For instance, [stocks: intc yhoo] will show information about Intel and Yahoo. (Note you must type the ticker symbols, not the company name.)
  • site:

    • If you include [site:] in your query, Google will restrict the results to those websites in the given domain. For instance, [help site:www.google.com] will find pages about help within www.google.com. [help site:com] will find pages about help within .com urls. Note there can be no space between the “site:” and the domain.
  • allintitle:

    • If you start a query with [allintitle:], Google will restrict the results to those with all of the query words in the title. For instance, [allintitle: google search] will return only documents that have both “google” and “search” in the title.
  • intitle:

    • If you include [intitle:] in your query, Google will restrict the results to documents containing that word in the title. For instance, [intitle:google search] will return documents that mention the word “google” in their title, and mention the word “search” anywhere in the document (title or no). Putting [intitle:] in front of every word in your query is equivalent to putting [allintitle:] at the front of your query: [intitle:google intitle:search] is the same as [allintitle: google search].
  • allinurl:

    • If you start a query with [allinurl:], Google will restrict the results to those with all of the query words in the url. For instance, [allinurl: google search] will return only documents that have both “google” and “search” in the url. Note that [allinurl:] works on words, not url components. In particular, it ignores punctuation. Thus, [allinurl: foo/bar] will restrict the results to page with the words “foo” and “bar” in the url, but won’t require that they be separated by a slash within that url, that they be adjacent, or that they be in that particular word order. There is currently no way to enforce these constraints.
  • inurl:

    • If you include [inurl:] in your query, Google will restrict the results to documents containing that word in the url. For instance, [inurl:google search] will return documents that mention the word “google” in their url, and mention the word “search” anywhere in the document (url or no). Putting “inurl:” in front of every word in your query is equivalent to putting “allinurl:” at the front of your query: [inurl:google inurl:search] is the same as [allinurl: google search].

Google search service is never intended to gain unauthorised access of data but nothing can be done if we ourselves kept data in the open and do not follow proper security mechanisms.

Essentially emails, username, passwords, financial data and etc. shouldn’t be available in public until and unless it’s meant to be. Example, our details with the bank are never expected to be available in a google search. But our social media details are available in public because we ourselves allowed it.

Ending Note

Google Search is very useful as well as equally harmful at the same time. Because it indexes everything available over the web.

You need to follow proper security mechanisms and prevent systems to expose sensitive data. Follow OWASP, it provides standard awareness document for developers and web application security.

Scraper API provides a proxy service designed for web scraping. With over 20 million residential IPs across 12 countries, as well as software that can handle JavaScript rendering and solving CAPTCHAs, you can quickly complete large scraping jobs without ever having to worry about being blocked by any servers.

Note: By no means Box Piper supports hacking. This article is written to provide relevant information only. Always adhering to Data Privacy and Security.

Any thoughts, let's discuss on twitter

Sharing this article is a great way to educate others like you just did.



If you’ve enjoyed this issue, do consider subscribing to my newsletter.


Subscribe to get more such interesting content !


Feel free to send honest and blunt feedback on how helpful this was to you or any suggestions/improvements.

Support my work and buy me a Coffee. It'll mean the world to me. 😇










To read more such interesting topics, let's go Home


Tech, Product, Money, Books, Life. Discover stuff, be inspired, and get ahead.
Box Piper on Twitter

More Blogs from the house of Box Piper: