40 comments

  • westont5 2 hours ago

    I'm not sure I've seen it mentioned yet that when Vercel rolled out their environment variable UI, there was no "sensitive" option https://github.com/vercel/vercel/discussions/4558#discussion.... There was ~2 years or more until it was introduced https://vercel.com/changelog/sensitive-environment-variables...

    • _pdp_ an hour ago

      Sensitive does not mean it is not readable. It is just simply not exposed through the UI. It can be easily leaked if you return a bit too much props from the action functions or routes.

      The only way to defend against these types of issues is to encrypt your environment with your own keys, with secrets possibly baked into source as there are no other facilities to separate them. An attacker would need to not only read the environments but also download the compiled functions and find the decryption keys.

      It is not ideal but it could work as a workaround.

      • harikb 44 minutes ago

        > with secrets possibly baked into source

        please don't suggest this. The right way is to have the creds fetched from a vault, which is programmed to release the creds auth-free to your VM (with machine level identify managed by the parent platform)

        This is how Google Secrets or AWS Vaults work.

        • jcgl 19 minutes ago

          > The right way is to have the creds fetched from a vault, which is programmed to release the creds auth-free to your VM

          Or have whatever deployment tool that currently populates the env vars instead use the same information to populate files on the filesystem (like mounting creds).

        • _pdp_ 23 minutes ago

          I was reffering to Vercel. Other cloud environments have much better mechanisms for securing secrets.

        • SoftTalker 27 minutes ago

          This is just another layer of indirection (which isn't bad; it adds to the difficulty of executing a breach). The fundamental problem with encrypted secrets is that at some point you need to access and decrypt them.

          • SAI_Peregrinus 7 minutes ago

            HSMs & similar can at least time-limit access to secrets to the period where an attacker can make requests to the HSM.

      • awestroke 7 minutes ago

        The better way defend against these types of issues is to avoid Vercel and similar providers

  • datadrivenangel 2 hours ago

    "Effective defense requires architectural change: treating OAuth apps as third‑party vendors, eliminating long‑lived platform secrets, and designing for the assumption of provider‑side compromise."

    Designing for provider-side compromise is very hard because that's the whole point of trust...

    • losvedir 2 hours ago

      As someone trying to think about OAuth apps at our SaaS, it certainly is very hard.

      Do any marketplaces have a good approach here? I know Cloudflare, after their similar Salesloft issue, has proposed proxying all 3rd party OAuth and API traffic through them. But that feels a little bit like trading one threat vector for another.

      Other than standard good practices like narrow scopes, shorter expirations, maybe OAuth Client secret rotation, etc, I'm not sure what else can be done. Maybe allowlisting IP addresses that the requests associated with a given client can come from?

      • mooreds an hour ago

        This was probably partly a Google refresh token theft (given the length of the access). No inside info, just looking at how the attack occurred.

        OAuth 2.1[0] (an RFC that has been around longer than I've been at my employer) recommends some protections around refresh tokens, either making them sender constrained (tied to the client application by public/private key cryptography) or one-time use with revocation if it is used multiple times.

        This is recommended for public clients, but I think makes sense for all clients.

        The first option is more difficult to implement, but is similar to the IP address solution you suggest. More robust though.

        The second option would have made this attack more difficult because the refresh token held by the legit client, context.ai, would have stopped working, presumably triggering someone to look into why and wonder if the tokens had been stolen.

        0: https://datatracker.ietf.org/doc/html/draft-ietf-oauth-v2-1

      • wouldbecouldbe an hour ago

        I mean the admin account had visibility of clients env vars, thats maybe not really great in the first place.

        • iririririr 23 minutes ago

          you'd think. but this is a js dev world.

          nextjs app bake all env vars on the client side code!! it's all public, unless you prefix the name with private_ or something.

    • nyc_data_geek1 an hour ago

      Corroborates that zero-trust until now has been largely marketing gibberish. Security by design means incorporating concepts such as these to not assume that your upstream providers will not be utterly owned in a supply chain attack.

  • _pdp_ an hour ago

    > OAuth trust relationship cascaded into a platform-wide exposure

    > The CEO publicly attributed the attacker's unusual velocity to AI

    > questions about detection-to-disclosure latency in platform breaches

    Typical! The main failures in my mind are:

    1. A user account with far too much privileges - possible many others like them

    2. No or limited 2FA or any form of ZeroTrust architecture

    3. Bad cyber security hygiene

    • JauntyHatAngle an hour ago

      Blaming AI is gonna be the security breach equivalent to blaming ddos when your website breaks isn't it.

      • anematode 29 minutes ago

        That part of his tweet made me laugh out loud. I don't understand who it's directed toward.

        • BoorishBears 15 minutes ago

          The market. Rauch is 'strategic' like that, he'd even use a moment like this sneak in a sound bite to froth the market he has so much skin in

          "Vercel CEO says AI accelerated attack on critical infrastructure"

  • saadn92 2 hours ago

    What bites people: rotating a vercel env variable doesn't invalidate old deployments, because previous deploys keep running with the old credential until you redeploy or delete them. So if you rotated your keys after the bulletin but didn't redeploy everything, then the compromised value is still live.

    Also worth checking your Google Workspace OAuth authorizations. Admin Console > Security > API Controls > Third-party app access. Guarantee there are apps in there you authorized for a demo two years ago that are still sitting with full email/drive access.

    • quentindanjou an hour ago

      Usually rotating a credential means that you invalidate the previous one. Never heard of rotating credentials that would only create new ones and keep the old ones active.

      • simlevesque 29 minutes ago

        But then every rotation would break production, wouldn't it ?

        • kstrauser 5 minutes ago

          Ideally, you can have a couple of working versions at any given time. For instance, an AWS IAM role can have 0 to 2 access keys configured at once. To rotate them, you deactivate all but one key, create a new key, and make that new key the new production value. Once everything's using that key, you can deactivate the old one.

    • wouldbecouldbe an hour ago

      When you rotate them, you supposed expire your old vars

    • kevinqi an hour ago

      yeah not redeploying on credential changes seems like a design flaw. Render redeploys on env var changes, for instance.

  • hungryhobbit 22 minutes ago

    Why is this same story repeated over and over here?

    I get it, it's a big story ... but that doesn't mean it needs N different articles describing the same thing (where N > 1).

    • jackconsidine 18 minutes ago

      New information here -- I had no idea that Env enumeration was happening MONTHS before the disclosure for example and that's part of why I come to HN.

      Would guess that double digit percent of readers have some level of skin in the game with Vercel

  • tom1337 an hour ago

    I still don't get how this exactly worked. Is the OAuth Token they talk about the one that you get when a user uses "Sign in with Google"? Aren't they then bound to the client id and client secret of that specific Google App the user signed in to? How were the attackers able to go from that to a control plane? Because even if the attacker knows the users OAuth token, the client id and the client secret, they can access the Google Drive etc. (which is bad, I get that) but I simply do not understand how they could log in into any Vercel systems from that point. Did they find the credentials in the google drive?

    • _pdp_ an hour ago

      Once you have a session token, which is what you get after you complete the oauth dance, you can issue requests to the API. It is simple as that. The minted token had permission to access the victim's inbox, most likely, which the attacker leveraged to read email and obtain one-time passwords, magic links and other forms of juicy information.

  • krooj an hour ago

    Interesting - I wonder if this isn't a case of theft on a refresh token that was minted by a non-confidential 3LO flow w/PKCE. That would explain how a leaked refresh token could then be used to obtain access, but does the Vercel A/S not implement any refresh token reuse detection? i.e.: you see the same R/T more than once, you nuke the entire session b/c it's assumed the R/T was compromised.

  • greenmilk 35 minutes ago

    To me the biggest (but not only) issue is that blindly connecting sensitive tools to 3rd party services has been normalized. Every time I hear the word "claw" I cringe...

  • akanet 31 minutes ago

    This article is solely overly wordy (probably ai) restatements of essentially just what vercel has publicly disclosed already

  • vaguemit 2 hours ago

    I recently went to BreachForums and the space was filled with this

  • throwaway27448 2 hours ago

    Do any services use vercel?

    • drusepth 2 hours ago

      It's a really common platform for vibe coded sites, as I understand it.

    • antonvs an hour ago

      Small startups often use it but typically outgrow it quickly unless they remain small and simple.

    • jdw64 an hour ago

      First of all, it is often used in Korea.

  • pphysch 2 hours ago

    Security-by-obfuscation is ridiculed but I'm a firm believer that preventing yourself from getting owned when someone is able to type 3 letters `env` is a worthy layer of defense. Even if those same secrets are unencrypted somewhere else on the same system, at least make them spend a bunch of time crawling through files and such.

    • Quarrelsome 2 hours ago

      It's ridiculed because its no protection on its own when an attacker is motivated. Its fine to add as an additional layer though if you want to make your space mildly custom to protect against broader attacks.

      I don't see how its necessarily relevant to this attack though. These guys were storing creds in clear and assuming actors within their network were "safe", weren't they?

      • pphysch 2 hours ago

        TFA cites "env var enumeration", likely implying someone got somewhere they shouldn't and typed 3 characters, as the critical attack that led to customers getting compromised.

        My point is sensitive secrets should literally never be exported into the process environment, they should be pulled directly into application memory from a file or secrets manager.

        It would still be a bad compromise either way, but you have a fighting chance of limiting the blast radius if you aren't serving secrets to attackers on an env platter, which could be the first three characters they type once establishing access.

        • lbarrow an hour ago

          I don't think that's what the attacker did here. Vercel is a PaaS product where other developers run apps. The enumerated environment variables were the env vars of Vercel's customers, which Vercel likely stores in a long-term data store. Rather than running `env` on a Linux box somewhere, the attacker may have just accessed that data store.