← back

Principled GraphQL

Integrity

One Graph

Ensuring that the graph is well-defined, stable, and consistent

Federated Implementation

Though there is only one graph, the implementation of that graph should be federated across multiple teams.

Track the Schema in a Registry

There should be a single source of truth for registering and tracking the graph.

Just like it's important to track source code in a version control system, it's important to track the definition of your graph in a schema registry. There should be a single schema registry for your company that is the authoritative definition of the graph, rather than relying on whatever processes are running at the moment or whatever code is checked in on a developer's laptop. Like a source control system, the schema registry should store the history of changes to the graph and who made them, and it should understand the concept of multiple versions of the graphs (for example, staging and production, or different development branches) in a way that parallels the software development process.

The schema registry should become the central hub of the system, powering developer tools, workflows, or any business processes that would benefit from awareness of the graph and any actual or proposed changes to it.

Agility

Rapidly rolling out the graph and continuously adapting it to changing needs

  1. Iteratively Improve Performance
  2. Use Graph Metadata to Empower Developers

Abstract, Demand-Oriented Schema

The schema should act as an abstraction layer that provides flexibility to consumers while hiding service implementation details.

Use an Agile Approach to Schema Development

It may be tempting to try to define, ahead of time, the “perfect schema” for all of your organization's data. Rather, what really makes a schema valuable is the degree to which it follows actual user requirements, which are never known perfectly up front and are constantly changing. The true path to the “perfect schema” is to make it easy for the graph to evolve in response to actual needs.

Fields shouldn't be added to the schema speculatively. Ideally, each field should be added only in response to a concrete need by a consumer for additional functionality

Updating the graph should be a continuous process. Rather than releasing a new “version” of the graph periodically, such as every 6 or 12 months, it should be possible to change the graph many times a day if necessary. New fields can be added at any time. To remove a field, it is first deprecated, and then removed when no consumers use it. The schema registry enables this agile evolution of the graph, together with processes and tooling that keep everyone aware of changes that could affect them. This ensures that only fully vetted changes can go into production.

Iteratively Improve Performance

Rather than optimizing every possible use of the graph, the focus should be on supporting the actual query shapes that are needed in production. Tooling should extract proposed new query shapes and surface them, before they go into production, to all affected service teams with latency requirements and projected query volume. Once the query is in production, its performance should be continuously monitored. If this principle is followed, problems should be easy to track back to the service that is not behaving as expected.

Use Graph Metadata to Empower Developers

Operations

Securely deploying the graph to production at scale

Authorization in a graph has two equally important aspects:

It is a mistake to allow users to perform any possible query regardless of cost, with no ability to manage its impact on production systems.

Best practices for demand control include:

Structured Logging

Capture structured logs of all graph operations and leverage them as the primary tool for understanding graph usage.

A wealth of information can be captured about each operation (read or write) that is performed on a graph: what user and app performed the operation, what fields were accessed, how the operation was actually executed, how it performed, and more. This information is highly valuable and should be systematically captured and made available for later use. Instead of a text log, it should be captured in a structured, machine readable format so that it can be leveraged for as many purposes as possible.

The record of a graph operation is called a trace. A trace should bring together all pertinent information about an operation in one place, including business information (who performed the operation, what was accessed or changed, which feature of which app built by which developer, whether it succeeded, how it performed) and purely technical information (which backend services were contacted, how each service contributed to latency, whether caches were used).

Because traces truly capture how a graph is being used, they can be used for a wide range of purposes:

Traces for all graph operations should be collected in one central place, so that there is one authoritative stream of traces. This stream can then be piped into other observability systems (perhaps after a simple transformation for existing systems that are not GraphQL-aware), or stored in one or more data warehouses for later use (aggregated and sampled as budget, use cases, and scale require).

Separate the GraphQL Layer from the Service Layer

Adopt a layered architecture with graph functionality broken into a separate tier rather than baked into every service.

Footnotes: https://principledgraphql.com