The newer global node IDs (which can be forced via the 'X-Github-Next-Global-ID' header [1]) have a prefix indicating the "type" of object delimited by an underscore, then a base64 encoded msgpack payload. For most objects it contains just a version (starting at 0) followed by the numeric "databaseId" field, but some are more complex.<p>For example, my GitHub user [2] has the node ID "U_kgDOAAhEkg". Users are "U_" and then the remaining data decodes to: [0, 541842] which matches the numeric ID for my user accepted by the REST API [3].<p>You shouldn't rely on any of this implementation of course, instead just directly query the "databaseId" field from the GraphQL API where you need interoperability. And in the other direction the REST API returns the "node_id" field for the GraphQL API.<p>For folks who finds this interesting, you might also like [4] which details GitHub's ETag implementation for the REST API.<p>[1] <a href="https://docs.github.com/en/graphql/guides/migrating-graphql-global-node-ids" rel="nofollow">https://docs.github.com/en/graphql/guides/migrating-graphql-...</a>
[2] <a href="https://api.github.com/user/541842" rel="nofollow">https://api.github.com/user/541842</a>
[3] <a href="https://gchq.github.io/CyberChef/#recipe=Find_/_Replace(%7B'option':'Regex','string':'%5E%5B%5E_%5D%2B_'%7D,'',true,false,true,false)From_Base64('A-Za-z0-9%2B/%3D',true,false)From_MessagePack()&input=VV9rZ0RPQUFoRWtn" rel="nofollow">https://gchq.github.io/CyberChef/#recipe=Find_/_Replace(%7B'...</a>
[4] <a href="https://github.com/bored-engineer/github-conditional-http-transport" rel="nofollow">https://github.com/bored-engineer/github-conditional-http-tr...</a>
> That repository ID (010:Repository2325298) had a clear structure: 010 is some type enum, followed by a colon, the word Repository, and then the database ID 2325298.<p>It's a classic length prefix. Repository has 10 chars, Tree has 4.
I wouldn't decode them like this, it's fragile, and global node IDs are <i>supposed</i> to be opaque in GraphQL.<p>I see that GitHub exposes a `databaseId` field on many of their types (like PullRequest) - is that what you're looking for? [1]<p>Most GraphQL APIs that serve objects that implement the Node interface just base-64-encode the type name and the database ID, but I definitely wouldn't rely on that always being the case. You can read more about global IDs in GraphQL in the spec in [2].<p>[1] <a href="https://docs.github.com/en/graphql/reference/objects#pullrequest" rel="nofollow">https://docs.github.com/en/graphql/reference/objects#pullreq...</a>
[2] <a href="https://graphql.org/learn/global-object-identification/" rel="nofollow">https://graphql.org/learn/global-object-identification/</a>
Also, as pointed out below, Github's GraphQL types also include fields like `permalink` and `url` (and interfaces like `UniformResourceLocatable`) that probably save you from needing to construct it yourself.
If you want to store metadata in identifiers, an easy fix to preventing users from depending on arbitrary characteristics is to encrypt the data. You see this a lot with pagination tokens.
I just want to point out that Opus 4.5 actually knows this trick and will write the code to decode the IDs if it is working with GitHub's API lol
The only GitHub identifier Ive ever bothered to store explicitly (I.e., in its own dedicated column) is an immutable URL key like issue/pr # or commit hash. I've <i>stored</i> comment ids but I've never thought about it. They just get sucked up with the rest of the JSON blob.<p>Not everything has to be forced through some normalizing layer. You can maintain coarse rows at the grain of each issue/PR and keep everything else in the blob. JSON is super fast. Unless you're making crosscutting queries along comment dimensions, I don't think this would ever show up on a profiler.
I remember a time when the v3 API didn't even have IDs and whenever someone in your org changed their username or renamed a repo you'd be left guessing who it is.<p>This is also the reason I wrote our current user onboarding / repo management code from scratch, because the terraform provider sucks and without any management you'll have a wave of "x got offboarded but they were the only admin on this repo" requests. Every repo is owned by a team. Access is only ever per-team.
> Every repo is owned by a team. Access is only ever per-team.<p>This is indeed the working pattern, and applies not just to GitHub and organizing teams there, it's a useful pattern to use everywhere. Don't think "Give access to this user" but rather "Give access to this team, which this user is current a part of" and it solves a lot of bothersome issues.
Hyrums law at its finest :D (or D: if you deeply care about correctness)
In database design typically it recommends giving out opaque natural keys, and keeping your monotonically increasing integer IDs secret and used internally.
That is a best practice for two real reasons:<p>1. You don't want third parties to know how many objects you have<p>2. You don't want folks to be able to iterate each object by incrementing the id<p>But if you have composite IDs like this, that doesn't matter. All objects that belong to a repository have the repository id inside them. Incrementing the id gives you more objects from the same repo. Incrementing the repo id gives you...a random object or nothing at all. And if your IDs include a little entropy or a timestamp, you've effectively kneecapped anyone who's trying to abuse this.
> You don't want folks to be able to iterate each object by incrementing the id<p>If you have a lot of public or semi-public data that you don't want people to page through, then I suppose this is true. But it's important to note that separate natural and primary keys are <i>not</i> a replacement for authorization. Random keys may mitigate an IDOR vulnerability but authorization is the correct solution. A sufficiently long and securely generated random token can be used as both as an ID and for authorization, like sharing a Google Doc with "anyone who has a link," but those requirements are important.
What if you used some id that does not allow to count objs like guid?
Maybe. Until your natural key changes. Which happens. A lot.<p>Exposing a surrogate / generated key that is effectively meaningless seems to be wise. Maybe internally Youtube has an index number for all their videos, but they expose a reasonably meaningless coded value to their consumers.
Most of what the author discovered is real and technically correct, but it is also undocumented, unsupported, and risky to rely on.<p>GitHub has changed node ID internals before, quietly. If they add a field to the MessagePack array, switch encodings, encrypt payloads, introduce UUID-backed IDs..<p>every system relying on this will break instantly.
Seems brittle unless GitHub is guaranteeing that they will never change their ID algorithms.
Github staff have racked up hundreds of contributions to Rails in recent years to extend the sharded/multiple database support, now you know why.
I had seen GitHub node IDs, although I had not used them or tried to decode them (although I could see they seem to be base64), since I only used the REST API, which reports node IDs but does not use them as input.<p>It looks like a good explanation of the node IDs, though. However, like another comment says, you should not rely on the format of node IDs.
Remember this article when you get upset that your own customers have come to rely on behavior that you told them explicitly not to rely on.<p>If it is possible to figure something out, your customers will eventually figure it out and rely on it.
Once a system has a sufficient number of users, it no longer matters what you "explicitly" promised in your documentation or contract.<p>Hyrum’s Law: all observable behaviors of your system will eventually be depended on by someone.<p>Even if you tell users not to rely on a specific side effect, once they discover it exists and find it useful, that behavior becomes an implicit part of your system's interface. As a result, engineers often find that "every change breaks someone’s workflow," even when that change is technically a bug fix or a performance improvement.<p>Reliance on unpromised behavior is something I was also introduced to as Kranz’s Law (or Scrappy's Law*), which asserts that things eventually get used for their inherent properties and effects, without regard for their intended purpose.<p>"I insisted SIGUSR1 and SIGUSR2 be invented for BSD. People were grabbing system signals to mean what they needed them to mean for IPC, so that (for example) some programs that segfaulted would not coredump because SIGSEGV had been hijacked.
This is a general principle — people will want to hijack any tools you build, so you have to design them to either be un-hijackable or to be hijacked cleanly. Those are your only choices."
—Ken Arnold in The Art Of Unix Programming
This makes no sense. I am developing a product in this space (<a href="https://codeinput.com" rel="nofollow">https://codeinput.com</a>) and GitHub API and GraphQl is a badly entangled mess but you don’t trick your way around ids.<p>There is actually a documented way to do it: <a href="https://docs.github.com/en/graphql/guides/using-global-node-ids" rel="nofollow">https://docs.github.com/en/graphql/guides/using-global-node-...</a><p>Same for urls, you are supposed to get them directly from GitHub not construct them yourself as format can change and then you find yourself playing a refactor cat-and-mouse game.<p>Best you can do is an hourly/daily cache for the values.
1. The list of "scopes" are the object hierarchy that owns the resource. That lets you figure out which shard a resource should be in. You want all the resources for the same repository on the same shard, otherwise if you simply hash the id, one shard going down takes down much of your service since everything is spread more or less uniformly across shards.<p>2. The object identifier is at the end. That should be strictly increasing, so all the resources for the same scope are ordered in the DB. This is one of the benefits of uuid7.<p>3. The first element is almost certainly a version. If you do a migration like this, you don't want to rule out doing it again. If you're packing bits, it's nearly impossible to know what's in the data without an identifier, so without the version you might not be able know whether the id is new or old.<p>Another commenter mentioned that you should encrypt this data. Hard pass! Decrypting each id is <i>decidedly</i> slower than b64 decode. Moreover, if you're picking apart IDs, you're relying on an interface that was never made for you. There's nothing sensitive in there: you're just setting yourself up for a possible (probable?) world of pain in the future. GitHub doesn't have to stop you from shooting your foot off.<p>Moreover, encrypting the contents of the ID makes them sort randomly. This is to be avoided: it means similar/related objects are not stored near each other, and you can't do simple range scans over your data.<p>You <i>could</i> decrypt the ids on the way in and store both the unencrypted and encrypted versions in the DB, but why? That's a lot of complexity, effort, and resources to stop randos on the Internet from relying on an internal, non-sensitive data format.<p>As for the old IDs that are still appearing, they are almost certainly:<p>1. Sharded by their own id (i.e., users are sharded by user id, not repo id), so you don't need additional information. Use something like rendezvous hashing to choose the shard.<p>2. Got sharded before the new id format was developed, and it's just not worth the trouble to change
> <i>GitHub's migration guide tells developers to treat the new IDs as opaque strings and treat them as references. However it was clear that there was some underlying structure to these IDs as we just saw with the bitmasking</i><p>Great, so now GitHub can't change the structure of their IDs without breaking this person's code. The lesson is that if you're designing an API and want an ID to be opaque you have to literally encrypt it. I find it really demoralizing as an API designer that I have to treat my API's consumers as adversaries who will knowingly and intentionally ignore guidance in the documentation like this.
> Great, so now GitHub can't change the structure of their IDs without breaking this person's code.<p>And that is all the fault of the person who treated a documented opaque value as if it has some specific structure.<p>> The lesson is that if you're designing an API and want an ID to be opaque you have to literally encrypt it.<p>The lesson is that you should stop caring about breaking people’s code who go against the documentation this way. When it breaks you shrug. Their code was always buggy and it just happened to be working for them until then. You are not their dad. You are not responsible for their misfortune.<p>> I find it really demoralizing as an API designer that I have to treat my API's consumers as adversaries who will knowingly and intentionally ignore guidance in the documentation like this.<p>You don’t have to.
Sounds like you’ve maybe never actually run a service or API library at scale. There’s so many factors that go into a decision like that at a company that it’s never so simple. Is the person impacted influential? You’ve got a reputation hit if they negatively blog about how you screwed them after something was working for years. Is a customer who’s worth 10% of your annual revenue impacted? Bet your ass your management chain won’t let you do a breaking change / revert any you made by declaring an incident.<p>Even in OSS land, you risk alienating the community you’ve built if they’re meaningfully impact. You only do this if the impact is minimal or you don’t care about alienating anyone using your software.
> The lesson is that you should stop caring about breaking people’s code who go against the documentation this way. When it breaks you shrug. Their code was always buggy and it just happened to be working for them until then. You are not their dad. You are not responsible for their misfortune.<p>Sure, but good luck running a business with that mindset.
You could also say, if I tell you something is an opaque identifier, and you introspect it, it's your problem if your code breaks. I told you not to do that.
I think more important than worrying about people treating an opaque value as structured data, is wondering _why_ they're doing so. In the case of this blog post, all they wanted to do was construct a URL, which required the integer database ID. Just make sure you expose what people need, so they don't need to go digging.<p>Other than that, I agree with what others are saying. If people rely on some undocumented aspect of your IDs, it's on them if that breaks.
Exposing what people need doesn’t guarantee that they won’t go digging. It is surprisingly common to discover that someone has come up with a hack that depends on implementation details to do something which you exposed directly and they just didn’t know about it.
GitHub actually do have both the database ID and URL available in the API:<p><a href="https://docs.github.com/en/graphql/reference/objects#pullrequest" rel="nofollow">https://docs.github.com/en/graphql/reference/objects#pullreq...</a><p>OP’s requirements changed and they hadn’t stored them during their crawl
> Great, so now GitHub can't change the structure of their IDs without breaking this person's code<p>OP can put the decoded IDs into a new column and ignore the structure in the future. The problem was presumably mass querying the Github API to get those numbers needed for functional URLs.
Literally how I designed all the public facing R2 tokens like multipart uploads. It’s also a security barrier because forging and stealing of said tokens is harder and any vulnerability has to be done with cooperation of your servers and can be quickly shut down if needed.
At a big enough scale, even your bugs have users
This is well understood - Hyrum's law.<p>You don't need encryption, a global_id database column with a randomly generated ID will do.
Hyrum's law is a real sonuvabitch.
Can GitHub change their API response rate? Can they increase it? If they do, they’ll break my code ‘cause it expects to receive responses at least after 1200ms. Any faster than that and I get race conditions. I selected the 1200ms number by measuring response rates.<p>No, you would call me a moron and tell me to go pound sand.<p>Weird systems were never supported to begin with.
The API contract doesn’t stipulate the behavior so GitHub is free to change as they please.
Who cares if their code is broken in this case? Stupid games stupid prizes.
> Somewhere in GitHub's codebase, there's an if-statement checking when a repository was created to decide which ID format to return.<p>I doubt it. That's the beauty of GraphQL — each object can store its ID however it wants, and the GraphQL layer encodes it in base64. Then when someone sends a request with a base64-encoded ID, there _might_ be an if-statement (or maybe it just does a lookup on the ID). If anything, the if-statement happens _after_ decoding the ID, not before encoding it.<p>There was never any if-statement that checked the time — before the migration, IDs were created only in the old format. After the migration, they were created in the new format.
Some commenters mentioned that the GraphQL API exposes the database IDs and a URL to the object directly in the schema. But the author did not know that and instead reverse-engineered the ID generation logic.<p>This is one of many reasons why GraphQL sucks. Developers will do anything to avoid reading docs. In the REST API, the developerId and url fields would be easily discoverable by looking at the API response. But in GraphQL, there is no way to get all fields. You need to get a list of fields from the docs and request them explicitly.
The author may well have been aware of this. However, since the author didn't retrieve those database IDs or URLs in the first place, they would have had to make further requests to retrieve them, which they wanted to avoid doing.<p>"I was looking at either backfilling millions of records or migrating our entire database, and neither sounded fun."