Skip to Content

Subscriptions

Hive Gateway fully supports federated subscriptions and behaves just like Federation GraphQL subscriptions in Apollo Router.

Subgraphs providing subscriptions can communicate with Hive Gateway through one of the following protocols:

Clients connecting to the Hive Gateway may use either:

When to use subscriptions

GraphQL subscriptions allows to keep the client updated in real time.

Most of the time, a PubSub system is used to propagate events in the backend system. A client can use subscriptions to receive those events, augmented with all the data it needs using the GraphQL ability to resolve additional fields.

Subscriptions can be used for applications that relies on events or live data, such as chats, IoT sensors, alerting, stock prices, etc…

Learn more about Subscriptions

Subscriptions in Gateways

In the context of a gateway, subscriptions are forwarded from the client to the subgraph implementing the subscribed field.

With the power of the Gateway, each events received from the upstream subgraph will be augmented with the requested data from other subgraphs, and then sent to the client.

The Hive Gateway also abstract away the underlying protocol used to transport the data. A client can use a different transport than the one used to connect with the upstream subgraph.

Configure subgraph transport

By default, Hive Gateway will always try to use the same transport for queries, mutations and subscriptions.

In the case of HTTP, the default is to protocol is GraphQL over SSE. We highly recommend it, since it’s the most performant and idiomatic.

If your subgraph doesn’t implement subscriptions over SSE, you can configure Hive Gateway to use GraphQL over WebSocket or HTTP Callback.

Whichever protocol is used by Hive Gateway to subscribe to the upstream subgraphs, downstream clients can subscribe to the gateway using any supported protocol.

Subscriptions using WebSockets

If your subgraph uses WebSockets for subscriptions support (like with Apollo Server), Hive Gateway will need additional configuration pointing to the WebSocket server path on the subgraph.

Please note that WebSocket for communications between Hive Gateway and subgraphs are suboptimal compared to other possible transports. We recommend using either SSE or HTTP Callbacks instead.

gateway.config.ts
import { defineConfig, type WSTransportOptions } from '@graphql-hive/gateway' export const gatewayConfig = defineConfig({ supergraph: 'supergraph.graphql', transportEntries: { // use "*.http" to apply options to all subgraphs with HTTP '*.http': { options: { subscriptions: { kind: 'ws', // override the path if it is different than normal http location: '/subscriptions' } } } } })

Subscriptions using HTTP Callback

If your subgraph uses HTTP Callback protocol for subscriptions, Hive Gateway will need additional configuration.

gateway.config.ts
import { defineConfig, type HTTPCallbackTransportOptions } from '@graphql-hive/gateway' export const gatewayConfig = defineConfig({ supergraph: 'supergraph.graphql', // Setup Hive Gateway to listen for webhook callbacks, and emit the payloads through PubSub engine webhooks: true, transportEntries: { // use "*.http" to apply options to all subgraphs with HTTP '*.http': { options: { subscriptions: { kind: 'http-callback', options: { // The gateway's public URL, which your subgraphs access, must include the path configured on the gateway. public_url: 'http://localhost:4000/callback', // The path of the router's callback endpoint path: '/callback', // Heartbeat interval to make sure the subgraph is still alive, and avoid hanging requests heartbeat_interval: 5000 } satisfies HTTPCallbackTransportOptions } } } } })

Subscriptions using mixed protocols

Hive Gateway supports using different transport for different subgraphs. By default, subscriptions will use the same transport than queries and mutation. This can be change using the transportEntries option.

The key of each entry determine which subgraph will be impacted:

  • *: all subgraphs
  • *.{transportKind}: all subgraphs using transportKind. For example, *.http will impact all subgraph using the http transport.
  • {subgraphName}: a specific subgraph.

Configuration are inherited and merged from the least specific to the most specific matcher. Only exception is the headers which is not inherited for the ws transport.

For example, let be 4 subgraphs:

  • products: using http transport for queries, and HTTP callbacks for subscriptions
  • views: using http transport for queries, and WS for subscriptions
  • stocks: using http transport for queries, and WS for subscriptions
  • stores: using mysql transport
gateway.config.ts
import { defineConfig, type WSTransportOptions } from '@graphql-hive/gateway' export const gatewayConfig = defineConfig({ transportEntries: { '*.http': { // Will be applied to products, views and stocks subgraphs, but not stores. options: { subscriptions: { kind: 'ws', options: { connectionParams: { token: '{context.headers.authorization}' } } satisfies WSTransportOptions } } }, products: { // Will override the subscriptions configuration for products subgraph only options: { subscriptions: { kind: 'http-callback', location: '/subscriptions', headers: [['authorization', 'context.headers.authorization']] } } } } })

Propagation of authentication and headers

Hive Gateway can propagate the downstream client’s Authorization header (or any other header) to the upstream subgraph.

The propagation of headers is different if you use pure HTTP transports (SSE or HTTP Callbacks) or WebSockets.

Propagation of headers for pure HTTP subscription transports follow the same configuration than normal upstream requests.

import { defineConfig } from '@graphql-hive/gateway' export const gatewayConfig = defineConfig({ propagateHeaders: { fromClientToSubgraphs({ request }) { return { authorization: request.headers.get('authorization') } } } })

Configure Client subscriptions

Client subscriptions are enabled by default and compatible with both GraphQL over SSE and GraphQL over WebSocket with graphql-ws.

The default endpoint for subscriptions is /graphql and follow the graphqlEndpoint option, as for queries and mutations.

You can disable WebSockets server by using disableWebsockets option in the config file or by providing --disable-websockets option to the hive-gateway CLI.

Closing active subscriptions on schema change

When the schema changes in Hive Gateway, all active subscriptions will be completed after emitting the following execution error:

{ "errors": [ { "message": "subscription has been closed due to a schema reload", "extensions": { "code": "SUBSCRIPTION_SCHEMA_RELOAD" } } ] }

This is also what Apollo Router when terminating subscriptions on schema update.

Example

We’ll implement two GraphQL Yoga federation services behaving as subgraphs. The “products” service exposes a subscription operation type for subscribing to product changes, while the “reviews” service simply exposes review stats about products.

The example is somewhat similar to Apollo’s documentation, except for that we use GraphQL Yoga here and significantly reduce the setup requirements.

You can find this example source on GitHub.

Install dependencies

npm i graphql-yoga @apollo/subgraph graphql

Services

In this example, we will compose 2 services:

  • Products which contains the products data
  • Reviews which contains reviews of products

Those 2 services needs to be run in parallel of the gateway.

products.ts
import { createServer } from 'http' import { parse } from 'graphql' import { createYoga } from 'graphql-yoga' import { buildSubgraphSchema } from '@apollo/subgraph' import { resolvers } from './my-resolvers' const typeDefs = parse(/* GraphQL */ ` type Product @key(fields: "id") { id: ID! name: String! price: Int! } type Subscription { productPriceChanged: Product! } `) const yoga = createYoga({ schema: buildSubgraphSchema([{ typeDefs, resolvers }]) }) const server = createServer(yoga) server.listen(40001, () => { console.log('Products subgraph ready at http://localhost:40001') })

Supergraph

Once all services have been started, we can generate a supergraph schema. It will then be served by the Hive Gateway.

You can generate this schema with either GraphQL Mesh or Apollo Rover.

To generate a supergraph with Apollo Rover, you first need to create a configuration file describing the list of subgraphs:

supergraph.yaml
federation_version: =2.3.2 subgraphs: products: routing_url: http://localhost:40001 schema: subgraph_url: http://localhost:40001 inventory: routing_url: http://localhost:40002 schema: subgraph_url: http://localhost:40002

You can then run the Rover command to generate the supergraph schema SDL:

rover supergraph compose --config ./supergraph.yaml > supergraph.graphql

For more details about how to use Apollo Rover, please refer to the official documentation.

Start Gateway

You can now start the Hive Gateway. Without any configuration provided, the Gateway will load the supergraph file supergraph.yaml from the current directory, and serve it with a set of sensible default features enabled.

hive-gateway supergraph

Subscribe

By default, subscriptions are enabled and handles both WebSockets and SSE transport.

Let’s now subscribe to the product price changes by executing the following query:

subscription { productPriceChanged { # Defined in Products subgraph name price reviews { # Defined in Reviews subgraph score } } }

Hive Gateway will inteligently resolve all fields on subscription events and deliver you the complete result.

You can subscribe to the gateway through Server-Sent Events (SSE) (in JavaScript, using EventSource or graphql-sse).

Most clients offers a way to use subscriptions over SSE. You can find here examples for Apollo Client and Relay, please refer to the Recipes for Clients Usage section of graphql-sse documentation for other clients setups.

To quickly test subscriptions, you can use curl in your terminal to subscribe to the gateway. curl has native support of SSE.

curl 'http://localhost:4000/graphql' \ -H 'accept: text/event-stream' \ -H 'content-type: application/json' \ --data-raw '{"query":"subscription OnProductPriceChanged { productPriceChanged { name price reviews { score } } }","operationName":"OnProductPriceChanged"}'

PubSub

Hive Gateway internally uses a PubSub system to handle subscriptions. By default, an in-memory PubSub engine is used when detecting subscriptions.

You can implement your own PubSub engine by implementing the PubSub interface from the @graphql-hive/pubsub package, which looks like this:

export type TopicDataMap = { [topic: string]: any /* data */ } export type PubSubListener<Data extends TopicDataMap, Topic extends keyof Data> = ( data: Data[Topic] ) => void type MaybePromise<T> = T | Promise<T> export interface PubSub<M extends TopicDataMap = TopicDataMap> { /** * Publish {@link data} for a {@link topic}. * @returns `void` or a `Promise` that resolves when the data has been successfully published */ publish<Topic extends keyof M>(topic: Topic, data: M[Topic]): MaybePromise<void> /** * A distinct list of all topics that are currently subscribed to. * Can be a promise to accomodate distributed systems where subscribers exist on other * locations and we need to know about all of them. */ subscribedTopics(): MaybePromise<Iterable<keyof M>> /** * Subscribe and listen to a {@link topic} receiving its data. * * If the {@link listener} is provided, it will be called whenever data is emitted for the {@link topic}, * * @returns an unsubscribe function or a `Promise<unsubscribe function>` that resolves when the subscription is successfully established. the unsubscribe function returns `void` or a `Promise` that resolves on successful unsubscribe and subscription cleanup * * If the {@link listener} is not provided, * * @returns an `AsyncIterable` that yields data for the given {@link topic} */ subscribe<Topic extends keyof M>(topic: Topic): AsyncIterable<M[Topic]> subscribe<Topic extends keyof M>( topic: Topic, listener: PubSubListener<M, Topic> ): MaybePromise<() => MaybePromise<void>> /** * Closes active subscriptions and disposes of all resources. Publishing and subscribing after disposal * is not possible and will throw an error if attempted. */ dispose(): MaybePromise<void> /** @see {@link dispose} */ [Symbol.asyncDispose](): Promise<void> }

The @grpahql-hive/pubsub package also provides a few built-in PubSub engines, at the moment an in-memory engine and a Redis-based engine.

In-Memory PubSub

The in-memory PubSub engine is the default engine used when subscriptions are detected. It can also be used explicitly by setting it in the configuration.

gateway.config.ts
import { defineConfig } from '@graphql-hive/gateway' import { MemPubSub } from '@graphql-hive/pubsub' // or from the Hive Gateway package import { MemPubSub } from '@graphql-hive/gateway' export const gatewayConfig = defineConfig({ supergraph: 'supergraph.graphql', pubsub: new MemPubSub() })

Redis PubSub

For more advanced use-cases, such as running multiple instances of Hive Gateway, you can use the Redis-based PubSub engine we offer out of the box.

In case you have distributed instances of Hive Gateway, using a distributed PubSub engine is required to make sure all instances are aware of all active subscriptions and can publish events to the correct subscribers.

For example, when using the webhooks transport for subscriptions, the subgraph will send events to only one instance of Hive Gateway. If that instance doesn’t have any active subscription for the topic, the event will be lost. Using a distributed PubSub engine solves this problem.

Redis PubSub does not come with Hive Gateway, you have to install the package and the Redis PubSub peer dependency of ioredis which you need to install first:

npm i @graphql-hive/pubsub ioredis
gateway.config.ts
import Redis from 'ioredis' import { defineConfig } from '@graphql-hive/gateway' import { RedisPubSub } from '@graphql-hive/pubsub/redis' /** * When a Redis connection enters "subscriber mode" (after calling SUBSCRIBE), it can only execute * subscriber commands (SUBSCRIBE, UNSUBSCRIBE, etc.). Meaning, it cannot execute other commands like PUBLISH. * To avoid this, we use two separate Redis clients: one for publishing and one for subscribing. */ const pub = new Redis() const sub = new Redis() export const gatewayConfig = defineConfig({ webhooks: true, pubsub: new RedisPubSub( { pub, sub }, { // we make sure to use the same prefix for all gateways to share the same channels and pubsub // meaning, all gateways using this channel prefix will receive and publish to the same topics channelPrefix: 'my-shared-gateways' } ) })

Now, with this setup, any instance of Hive Gateway using the same channelPrefix will be able to share the same subscriptions.

💡

Note that this works only if you have subgraphs that publish subscriptions to Hive Gateway’s webhook to the same pubsub topic (see documentation on using webhooks to handle subscriptions).

You should also take a look at the E2E test serving as an example of how distributed subscriptions would work with multiple Hive Gateway instances.

Last updated on