99 comments

  • landl0rd 2 days ago

    The Cloudflare Radar page is probably a more representative sample: https://radar.cloudflare.com/domains

  • GCUMstlyHarmls 2 days ago

    transmissionbt.com (A bittorrent client for macOS) is out ranking youtube, wikipedia, github, etc. Is transmission that popular? I assume its the auto-updater? Seems insane.

    https://github.com/Quad9DNS/quad9-domains-top500/blob/dfd513...

  • 1vuio0pswjnm7 2 days ago

    As the comments here suggest, this list may be more indicative of some developer-introduced application behaviour, e.g., gratuitous DNS lookups, than "popularity".

  • cbsks 2 days ago

    What’s up with wvdbozpfc.com?

    There’s a bunch of random looking domain names: cmidphnvq.com, rpqihexdb.com, facebook.com. I’d guess they for advertising?

    • BearOso 2 days ago

      I looked up a couple. They're cloudflare regional servers.

    • t0mas88 2 days ago

      Or malware, those would typically be fairly random domain names that are queried for updates or instructions by a large number of infected devices.

      • 0points 2 days ago

        That's what I'm thinking too. That would suggest some very large operational botnets ... :-/

        • netsharc 2 days ago

          Or they query the DNS very often. Most devices have DNS caching, so if things like tiktok.com end up there, there must be a loot of devices (also, a lot of subdomains, which aren't visible in these lists).

    • mammuth 2 days ago

      Are there host lists for pihole/adguard/ublock for these kinds of domains?

      I'd assume the domains change regularly if it's malware or bot networks, but because they rank so high in this list, it sounds like it should be feasible to keep a blocklist somewhat up to date.

      • danudey 2 days ago

        It could also be ad networks; create random domains and subdomains so that simple domain blocklists are difficult to keep up to date efficiently (or at least, so that constant maintenance is required).

      • homebrewer 2 days ago

        https://gitlab.com/malware-filter

        Some of these lists are already in uBO out of the box.

    • gmuslera 2 days ago

      It could be a good pattern for spam/ads organizations, changing the random domain name as soon as traffic drops because the actual ones ended in enough blocklists.

    • forty 2 days ago

      Also blockdh100b ?

      • maxmcd 2 days ago

        router.blockdh100b.net resolves

        so does router.blockdh100c.co

  • supriyo-biswas 2 days ago

    Seems like it'd be a good addition to the Tranco list: https://tranco-list.eu/

    • joelesler 2 days ago

      They already are. Source: DNS Researcher myself.

    • peterdavehello 2 days ago

      In addition to Tranco, I maintain regularly updated lists of the top one million domains from sources like Cisco, Majestic, BuiltWith, Statvoo, DomCop, and Cloudflare. Feel free to check it out: https://github.com/PeterDaveHello/top-1m-domains

    • tptacek 2 days ago

      Came here to say: if people are interested in this stuff, they should just pull down the Tranco list --- it includes feeds from Quad9 and Cloudflare.

  • jeffbee 2 days ago

    I don't see how it would be possible to produce this table under Quad9's privacy policy. Nothing in their privacy policy says that they maintain logs that would enable them to count queries by label. Can anyone explain?

    • ratorx 2 days ago

      It does say that they collect this information in their “Data and Privacy Policy”. Specifically section 2.2 (Data Collected): https://quad9.net/privacy/policy/

      Which policy are you referring to that implies they don’t?

      Also I think you are assuming they store query logs and then aggregate this data later. It is much simpler just to maintain an integer counter for monitoring as the queries come in, and ingest that into a time series database (not sure if that’s what they actually do). Maybe it needs to be a bit fancier to handle the cardinality of DNS names dimension, but re-constructing this from logs would be much more expensive.

      • yawndex 2 days ago

        The section you mentioned does not say anything about having counters for labels. It only mentions that they record "[t]he times of the first and most recent instances of queries for each query label".

        • rpdillon 2 days ago

          Well, the counters aren't data collected, they are data derived from the data they do collect. The privacy policy covers collection.

          EDIT: I see they went out of their way to say "this is the complete list of everything we count" and they did not include counters by label, so I see your point!

      • jeffbee 2 days ago

        I don't see how that is compatible with 2.2. They don't say anything about counters per label. It says counter per RR type, and watermarks of least and most recent timestamps by label, not count by label.

        If an organization is going to be this specific about what they count, it implies that this is everything they count, not that there may also be other junk unmentioned.

    • danudey 2 days ago

      I took a look at their privacy policy and agree that it doesn't specifically list that it logs which domains are being queried. It does list a bunch of things it does log as counters, all of which seems reasonable, but they don't explicitly say "we count which domains are being queried".

      That said, I think it's entirely reasonable for them to log domains alone if they're completely disconnected from any user activity, i.e. a simple "increment the counter for foo.com" is reasonable since that's unrelated to user privacy.

      • staplers 2 days ago

        Unless say, an adversary can link an obscure domain to a specific user/use case. Get that counter log and you can track a certain behavior (only pings this domain when about to do something or when on vacation, their house is empty, etc.)

        • Bender 2 days ago

          One way around that is to set up a cron job that queries the most common domains one visits hourly. When requested by workstations and cell phones they will be served up by cache. At least that is what I have been doing for a few decades and works fine. I block all the DoH/DoT resolvers which is easier to do than some might think. One can do the individual A records or just the apex A/NS records to get infrastructure cache and then configure Unbound to prefetch records about to expire.

          Just for fun I have added some of these into my cron job.

        • Tijdreiziger 2 days ago

          The average burglar probably isn’t cross-referencing DNS statistics.

          • staplers 2 days ago

            Yes but I don't think the average Quad9 user is worried about being tracked by average burglars.. more like advanced syndicates and nation states.

            • natebc 2 days ago

              You think the average quad9 user is being tracked by advanced syndicates and nation states?

              • greyface- 2 days ago

                Such surveillance is often done on a dragnet basis, so, yes.

            • Tijdreiziger 2 days ago

              Huh? The average Quad9 user is probably a tech-savvy person who cares about online privacy and/or malware protection (Quad9 blocks known malicious domains).

  • onlyrealcuzzo 2 days ago

    It's quite interesting to me that ChatGPT is in the 200s and 300s.

    By almost every metric this is one of the 10 busiest websites, and some sources are already putting it in the top 5.

    Are they just disproportionately not using Quad9?

    I understand that there's a lot of overlap with Google having several spots in the top 50 itself, several being infrastructure like cloudflare and akamai, and several others being malware - but it still seems surprising.

    It's just kind of shocking to see Slack, Zoom, LinkedIn, and even DropBox, Roku, and Yandex much higher up.

    • Bender 2 days ago

      Something else to factor in is the TTL of both NS/A types for each apex domain and the individual records including sub-domains. Clients will not be querying Quad9 until the TTL expires on their clients. TTL would have to be factored into query rates to determine popularity correctly whereas these lists just show raw query numbers.

      For example, there are many records under amazonaws.com that have 5 second TTL's mostly EC2 instances. As such clients will query them at a much higher rate whereas grammarly.io have a number of records with a 900 second TTL. This will skew the ranking positions of the two apex domains. I suppose if one wanted to game this they could have an A record to a non-critical part of a site that is not visibly rendered by the end-user and has a TTL of 1 second assuming quad9 is not rewrite min/max-ttl which some resolvers do.

      Examples of just some of the TTL's used on these apex domains excluding individual records:

          30 32 60 300 600 900 1200 1800 3600 7200 10800 21600 28800 43200 86400 90000 3600000
      
      Some examples of rewriting max-ttl I forgot which ones rewrite min-ttl:

          for Resolver in 1.1.1.1 8.8.8.8 9.9.9.9 216.128.176.142;do echo -en "${Resolver}:\t"; dig @${Resolver} +nocookie +noall +answer -t a big.ohcdn.net;done | column -t
          1.1.1.1:          big.ohcdn.net.  3628800  IN  A  227.227.227.227
          8.8.8.8:          big.ohcdn.net.  21422    IN  A  227.227.227.227
          9.9.9.9:          big.ohcdn.net.  43200    IN  A  227.227.227.227
          216.128.176.142:  big.ohcdn.net.  3628800  IN  A  227.227.227.227  # authoritative server
      
      [Edit] I just realized they made a general statement to this effect in the git repo.
    • QuinnyPig 2 days ago

      My theory: the domains you name have ad beacons, desktop apps that are persistently running, and/or physical devices plugged into networks out there. Whereas ChatGPT is used (domainwise) overwhelmingly by humans hitting the site in their browsers.

    • danielcid 2 days ago

      Mostly because of sub domains. They are counting all the sub domains requests to give the top domains ranking.

      Some of those have many trackers and background sub domains that add up.

      For example, Linkedin their most popular sub domain is: px.ads.linkedin.com

      Here is a more comprehensive list with top 10k domains (including sub domains):

      https://dnsarchive.net/top-domains?rank=top10k

    • doubleorseven 2 days ago

      i also looked it up. it feels like up until ~200 those are all just chatty apps on our computers talking with the mother ship

  • jtbayly 2 days ago

    I expected to see porn in the list.

    • atomicnumber3 2 days ago

      I was personally going to be surprised. Bots and machines categorically do not peruse such material, and DNS traffic is largely not going to have a human on the other end.

  • trutz 2 days ago

    What is amazon.dev? Does not resolve for me.

    {"position": 127, "domain_name": "amazon.dev", "date": "2025-07-10"}

    Source: https://github.com/Quad9DNS/quad9-domains-top500/blob/main/t...

    • jvolkman 2 days ago

      https://www.google.com/search?q=%22amazon.dev%22

      Looks like their customer support rep portal. Presumably there are not A/CNAME records at the top level, but na.headphones.whs.amazon.dev resolves.

    • angott 2 days ago

      I’m not entirely sure what it is, but my Alexa devices hit subdomains within it very frequently based on my local DNS history. That’s probably why it made the top of the list.

  • mekster 2 days ago

    Who are looking up PTR records?

    54.in-addr.arpa looks to be Amazon's range and there are several others.

    • mccoyc 2 days ago

      It's probably a lot of automated tooling/monitoring infrastructure that's doing reverse resolution of IPs to get hostnames.

      Edit: I've found that sometimes they're pretty poor at caching responses so you end up with a lot of these requests.

      • citrin_ru 2 days ago

        Mail servers typically resolve a remote IP to a PTR. High number of PTR requests can indicate that the network is used to send email. Amazon (both SES and EC2) is one of the biggest email sources on the Internet (ranging from ham to marketing and there is huge spam volume from AWS too).

        Thought I don’t expect mail servers to use quad9.

        • mekster a day ago

          Why not?

          I have unbound with upstream set to 1.1.1.1 and 9.9.9.9.

          • citrin_ru a day ago

            If you run a mail server it's good to have a local DNS cache, e. g. Unbound. And if you have Unbound adding an upstream is not that necessary - it can talk to root servers directly or use hoster's DNS as upstream. I rarely use services like Quad9 but IMHO the main use case is for home PC / notebooks: e. g. if an ISP neglected it's own DNS (and it doesn't work well) or if an ISP spoof NXDOMAIN to return ads; another use case is free Wi-Fi where DNS can also be misconfigured / unreliable.

            For an ISP it's relatively easy to provide a DNS server which will be fast and reliable (and your ISP's DNS is close to you than some 3rd party DNS) if that's not the case they probably just don't care.

  • efitz 2 days ago

    shodan.io > gmail.com on 6/1 ???

  • mekster 2 days ago

    Isn't part of the reasons to run a public DNS to sell these hard earned info for profit to marketers etc but they just release publicly? Of course this is just the tip of the iceberg of the information they gather.

    Really interesting to know though.

    Some just look way high up and could mean buggy implementation without proper cache usage or persistently banging the domain.

    • null0ranje 2 days ago

      Data in the aggregate is quite interesting and probably of little value to marketers.

  • wigster 2 days ago

    example.com #17 ?

    • mekster 2 days ago

      This just shows that domain is incorporated not just in documents but on systems that actually trigger accessing it all over the world.

    • paweladamczuk 2 days ago

      Funny, people don't bother reading RFCs

    • SoftTalker 2 days ago

      Vibe coders.

  • xhrpost 2 days ago

    Has anyone used Quad9 and also NextDNS and have thoughts on how they compare?

  • 0points 2 days ago

    > https://github.com/Quad9DNS/quad9-domains-top500/blob/main/t...

    {"position": 5, "domain_name": "kxulsrwcq.com", "date": "2025-07-10"}

    What the

    https://www.ipaddress.com/website/kxulsrwcq.com/

    > Safety/Trust: Unknown

    • Matheus28 2 days ago

      Probably some sort of command and control for a botnet.

      They calculate a random domain name based on the timestamp (so it’s constantly changing every X days in case it gets seized), and have some validation to make sure commands are signed (to prevent someone name squatting to control their botnet).

      • threeducks 2 days ago

        Wow, that's smart. I was wondering whether there is a way for the bots to generate "unpredictable" domains such that security researchers could not predict them efficiently (even with source code), but the botnet controller can.

        Time-lock puzzles come close, but but it requires that the bots have computing power comparable to the security researchers.

        • orlp 2 days ago

          > Wow, that's smart. I was wondering whether there is a way for the bots to generate "unpredictable" domains such that security researchers could not predict them efficiently (even with source code), but the botnet controller can.

          There is a fairly simple method which achieves the same advantage for a botnet controller.

          1. Use a hash of the current day to derive, for that day, an infinite stream of domain names. This could be something as simple as `to_human_readable_domain(sha256(daily_hash + i))`.

          2. A botnet slave attempts to access servers in a diagonal order over (days, domains), starting at the first domain for today and working backwards in days and forwards in domains. An image best describes what I mean by this: https://i.imgur.com/lcEbHwz.png

          3. So long as one of those domains is controlled by the botnet operator (which can be verified using a signed response from the server), they can control the botnet.

          This means that the botnet operator only needs to purchase one domain every couple of days to keep controlling their botnet, while someone trying to stop them will have to buy thousands and thousands every day.

          And when you successfully purchase a domain you can publish the new domain to any connected slaves, so this scheme is only necessary for recruitment into the network, not continued control.

          • tea-lover 2 days ago

            Here's the same image on a less horrible file hosting:

            https://files.catbox.moe/gilmd1.png

            Imgur has been inaccessible for me for months, they're one of those organizations that consider it proper to block whole countries to counter bot abuse.

            • stirfish 2 days ago

              Hmm, catbox used to be blocked for me too, but I can access it today. That's interesting.

          • itake 2 days ago

            I've definitely heard of cnc using a plural of domains for this reason. the bots have a list of domains they reach out to, searching for one that is valid.

            I believe one issue with this strategy is many corporate VPNs block fresh domains. I guess if the software was pinned to use encrypted DNS instead of whatever the OS recommends, then the DNS blocking could be avoided...

            • m-s-y 2 days ago

              How would a corporate DNS block new domains, exactly?

              • itake 2 days ago

                My employer uses Zscaler. I don't know exactly how they implement this, but my educated guess is the corporate DNS server doesn't resolve domains that were created recently.

                In technical terms, the device asks the private corporate DNS server for the IP address of the hostname. The private DNS server checks the requested domain against a threat intelligence feed that tracks domain registration dates (and security risks). If the domain is deemed a threat, either return an IP address which points at a server that shows a warning message (if http traffic) or return an invalid IP (0.0.0.0).

              • aaronax 2 days ago

                A firewall. For example, Palo Alto firewalls can easily be configured to block domains newer than ~30 days old.

                https://knowledgebase.paloaltonetworks.com/KCSArticleDetail?...

              • paranoidrobot 2 days ago

                Have a cache of domains you know about with registration date.

                When getting a query for a domain you have not heard about, query whois for it. Store it's registration date in the cache.

        • sim7c00 2 days ago

          there are tools pretty good at detecting DGAs these days, but not often implemented.

          the best thing to do afaik is use services normal user shave access to, and communicate via those. its hard to tell for anyone who's extracting the data from the third party so the server is hidden. (e.g bot posts images to twitter, and server scrapes the images from twitter, this is also already old news but easier and more likely to sail through that next gen firewall -_-)

          i'd say having ur 'own' servers and domains is maybe even a bit dated ( though sadly still very effective!)

          • immibis 2 days ago

            It's one of many possible strategies. Any one strategy can be blocked if it's used by enough malicious actors (e.g. Twitter can be forced to block base64 tweets); if they all use different strategies, it becomes harder to justify blocking each individual one.

            • sim7c00 2 days ago

              you either need whitelisting, which ppl dont want because they need to send tweets and sync gdrive on their corpo laptops ;')...

              so i guess that leaves u with modeling normal user behavior to spot anomalies without the actual packet data being an indicator.

              then the bots could piggyback on regular coms still, but it'd definitely raise the bar...

        • afandian 2 days ago

          I can see a future where Cloudflare or similar offer a DNS + proxy + Root CA combo to intercept these. Maybe they already do.

          • lurkshark 2 days ago

            If I’m remembering correctly, Conficker was the first major use of this technique. They used a relatively small domain pool (250) so the registries were able to lock them up preemptively.

            I remember a couple legitimate sites getting slammed by accidental DDOS because the algorithm happened to generate their domain, but having a hard time finding a reference to that.

            https://en.m.wikipedia.org/wiki/Conficker

          • Tijdreiziger 2 days ago

            Quad9 (the subject of this post) already offers ‘threat blocking’ by default.

            https://quad9.net/service/threat-blocking/

          • threeducks 2 days ago

            That might work for the current generation of bots, but it will become infeasible when the domain names are generated in such a way that they overlap with spellable and existing domain names.

            • 0points 2 days ago

              > it will become infeasible when the domain names are generated in such a way that they overlap with spellable and existing domain names.

              And why do you believe this will even happen?

        • 2 days ago
          [deleted]
        • bobmcnamara 2 days ago

          Use a hash chain!

          Each time you resolve, the resulting IP can be part of the hash for predicting a future hostname.

    • 0points 2 days ago

      More:

      {"position": 26, "domain_name": "cmidphnvq.com", "date": "2025-07-10"}

      {"position": 28, "domain_name": "xmqkychtb.com", "date": "2025-07-10"}

      {"position": 37, "domain_name": "ezdrtpvsa.com", "date": "2025-07-10"}

      {"position": 38, "domain_name": "wvdbozpfc.com", "date": "2025-07-10"}

      {"position": 46, "domain_name": "bldrdoc.gov", "date": "2025-07-10"}

      {"position": 52, "domain_name": "gadf99632rm.xyz", "date": "2025-07-10"}

      • netsharc 2 days ago

        bldrdoc.gov seems to be Cisco devices looking for a time server: https://community.cisco.com/t5/ipv6/cisco-switch-generating-...

        Geniuses...

        • 0points 2 days ago

          Oh, hah. Well that doesn't seem intentionally malicious then.

          I added it in the first place as it was a non-resolving .gov in the top 50 list which seemed out of place to me.

          > bldrdoc.gov: No address associated with hostname

          I see that the time related subdomains in your link do resolve to the nist.gov timeserver.

          But I really am wondering what's up with all of the rest of these domains.

          • netsharc 2 days ago

            Googling it, one of the results was https://library.bldrdoc.gov/ , for Boulder Labs Library.

            More googling gave me https://www.boulder.doc.gov

            > Boulder is the home of scientific laboratories for the U. S. Department of Commerce’s NOAA, NIST and NTIA. Clustered on the foothills of the Rocky Mountains in Boulder Colorado, these labs are the home of scientific research and engineering in the fields of electromagnetics, materials reliability, optoelectronics, quantum electronics and physics, time and frequency, earth systems, weather and telecommunications.

            Looks like a place full of scientific knowledge. I hope they haven't suffered much DOGEing.

    • miyuru 2 days ago

      google the domains and you will find subdomains that point to cachefly.

          hiwd.kxulsrwcq.com is pointing to vdd.cachefly.net 
      
      I am not sure, but my guess is they might be used by some kind of a streaming service.
      • danudey 2 days ago

        Most likely something like an ad service to prevent their content being caught by domain blocklists. That would be similar to how a lot of websites started using randomized strings for attributes like id and class so that users couldn't block page elements based on CSS selectors.

        • gchamonlive 2 days ago

          Interesting how ad services and botnets behave similarly in some aspects

          • mlinhares 2 days ago

            Cue in "Are we the baddies?" meme.

          • immibis 2 days ago

            They are both malicious software that lots of people want to block; one happens to be legal.

    • danielcid 2 days ago

      And they are often used with random sub domains as well (but they did not include sub domains in their list).

      Ex:

      https://dnsarchive.net/search?q=cmidphnvq.com

      https://dnsarchive.net/search?q=xmqkychtb

      https://dnsarchive.net/ipv4/34.126.227.30

    • whalesalad 2 days ago

      One of the CNAME's defined for that domain is "hiwd.kxulsrwcq.com" which appears to be flagged for malware. https://www.securefeed.com/Content/WebLookup?host=hiwd.kxuls...

    • reactordev 2 days ago
    • 2 days ago
      [deleted]