Realtime
One of the most often-asked questions of Redwood before and after the launch of V1 was, “When will Redwood support a realtime solution?”
The answer is: now.
What is Realtime?
Redwood's initial realtime solution leverages GraphQL and relies on a serverful deployment to maintain a long-running connection between the client and server.
This means that your cannot use Realtime when deploying to Netlify or Vercel.
See one of Redwood's many other Deploy providers, and the Docker setup for good measure.
Redwood's GraphQL server uses the GraphQL over Server-Sent Events spec's "distinct connections mode" for subscriptions.
Advantages of SSE over WebSockets include:
- Transported over simple HTTP instead of a custom protocol
- Built in support for re-connection and event-id
- Simpler protocol
- No trouble with corporate firewalls doing packet inspection
Subscriptions and Live Queries
In GraphQL, there are two options for real-time updates: live queries and subscriptions.
Subscriptions are part of the GraphQL specification, whereas live queries are not.
There are times where subscriptions are well-suited for a realtime problem and in some cases live queries may be a better fit. Later we’ll explore the pros and cons of each approach and how best to decide which to use and when.
Defer and Stream
Defer and stream are directives that allow you to improve latency for clients by sending the most important data as soon as it's ready.
As applications grow, the GraphQL operation documents can get bigger. The server will only send the response back once all the data requested in the query is ready. But not all requested data is of equal importance, and the client may not need all of the data at once.
Using Defer
The @defer
directive allows you to postpone the delivery of one or more (slow) fields grouped in an inlined or spread fragment.
Using Stream
The @stream
directive allows you to stream the individual items of a field of the list type as the items are available.
The @stream
directive is currently not supported by Apollo GraphQL client.
Features
Realtime handles the hard parts of a GraphQL realtime implementation by automatically:
- allowing GraphQL Subscription operations to be handled
- merging in your subscriptions types and mapping their handler functions (subscribe and resolve) to your GraphQL schema letting you keep your subscription logic organized and apart from services (your subscription may use a service to respond to an event)
- authenticating subscription requests using the same
@requireAuth
directives already protecting other queries and mutations (or you can implement your own validator directive) - adding in the
@live
query directive to your GraphQL schema and setting up theuseLiveQuery
envelop plugin to handle requests, invalidation, and managing the storage mechanism needed - creating and configuring in-memory and persisted Redis stores used by the PubSub transport for subscriptions and Live Queries (and letting you switch between them in development and production)
- placing the pubSub transport and stores into the GraphQL context so you can use them in services, subscription resolvers, or elsewhere (like a webhook, function, or job) to publish an event or invalidate data
- typing your subscription channel event payloads
- support
@defer
and@stream
directives
It provides a first-class developer experience for real-time updates with GraphQL so you can easily
- respond to an event (e.g. NewPost, NewUserNotification)
- respond to a data change (e.g. Post 123's title updated)
and have the latest data reflected in your app.
Lastly, the Redwood CLI has commands to generate a boilerplate implementation and sample code needed to create your custom subscriptions and Live Queries.
Regardless of the implementation chosen, a stateful server and store are needed to track changes, invalidation, and who wants to be informed about changes.
What can I build with Realtime?
- Application alerts and messages
- User notifications
- Live charts
- Location updates
- Auction bid updates
- Messaging
- OpenAI streaming responses
Redwood Realtime Setup
To setup realtime in an existing Redwood project, run the following commands:
yarn rw setup server-file
yarn rw setup realtime
You'll get:
api/server.ts
where you can configure your Fastify serverapi/lib/realtime.ts
where you consume your subscriptions and configure realtime with an in-memory or Redis store- Usage examples for live queries, subscriptions, defer, and stream. You'll get sdl, services/subscriptions for each
- The
auction
live query example - The
countdown timer
subscription example - The
chat
subscription examples - The
alphabet
stream example - The
slow and fast
field defer example
There is no UI set up for these examples. You can find information on how to try them out using the GraphiQL playground.
Just add the realtime configuration to your GraphQL handler in api/src/functions/graphql.ts
and you're good to go:
+ import { realtime } from 'src/lib/realtime'
export const handler = createGraphQLHandler({
// ...
+ realtime,
})
Realtime Configuration
By default, Redwood's realtime configures an in-memory store for the Pub Sub client used with subscriptions and live query invalidation.
Realtime supports in-memory and Redis stores:
- In-memory stores are useful for development and testing.
- Redis stores are useful for production.
To enable defer and streaming, set enableDeferStream
to true.
Configure a Redis store and defer and stream in:
import { RedwoodRealtimeOptions } from '@redwoodjs/realtime'
import subscriptions from 'src/subscriptions/**/*.{js,ts}'
// if using a Redis store
// import { Redis } from 'ioredis'
// const publishClient = new Redis()
// const subscribeClient = new Redis()
/**
* Configure RedwoodJS Realtime
*
* See https://redwoodjs.com/docs/realtime
*
* Realtime supports Live Queries and Subscriptions over GraphQL SSE.
*
* Live Queries are GraphQL queries that are automatically re-run when the data they depend on changes.
*
* Subscriptions are GraphQL queries that are run when a client subscribes to a channel.
*
* Redwood Realtime
* - uses a publish/subscribe model to broadcast data to clients.
* - uses a store to persist Live Query and Subscription data.
*
* Redwood Realtime supports in-memory and Redis stores:
* - In-memory stores are useful for development and testing.
* - Redis stores are useful for production.
*/
export const realtime: RedwoodRealtimeOptions = {
subscriptions: {
subscriptions,
store: 'in-memory',
// if using a Redis store
// store: { redis: { publishClient, subscribeClient } },
},
liveQueries: {
store: 'in-memory',
// if using a Redis store
// store: { redis: { publishClient, subscribeClient } },
},
// To enable defer and streaming, set to true.
// enableDeferStream: true,
}
PubSub and LiveQueryStore
By setting up realtime, the GraphQL server adds two helpers on the context:
- pubSub
- liveQueryStory
With context.pubSub
you can subscribe to and publish messages via context.pubSub.publish('the-topic', id, id2)
.
With context.liveQueryStore.
you can context.liveQueryStore.invalidate(key)
where your key may be a reference or schema coordinate:
Reference
Where the query is: auction(id: ID!): Auction @requireAuth
:
"Auction:123"
Schema Coordinate
When the query is: auctions: [Auction!]! @requireAuth
:
"Query.auctions"
Subscriptions
Redwood has a first-class developer experience for GraphQL subscriptions.
Subscribe to Events
- Granular information on what data changed
- Why has the data changed?
- Spec compliant
Chat/New Message Example
type Subscription {
newMessage(roomId: ID!): Message! @requireAuth
}
- I subscribed to a "newMessage” in room “2”
- Someone added a message to room “2” with a from and body
- A "NewMessage" event to Room 2 gets published
- I find out and see who the message is from and what they messaged (the body)
Countdown Timer Example
Counts down from a starting values by an interval.
subscription CountdownFromInterval {
countdown(from: 100, interval: 10)
}
This example showcases how a subscription yields its own response.
Live Queries
Redwood has made it super easy to add live queries to your GraphQL server! You can push new data to your clients automatically once the data selected by a GraphQL operation becomes stale by annotating your query operation with the @live
directive.
The invalidation mechanism is based on GraphQL ID fields and schema coordinates. Once a query operation has been invalidated, the query is re-executed, and the result is pushed to the client.
Listen for Data Changes
- I'm not interested in what exactly changed it.
- Just give me the data.
- This is not part of the GraphQL specification.
- There can be multiple root fields.
Auction Live Query Example
query GetCurrentAuctionBids @live {
auction(id: "1") {
bids {
amount
}
highestBid {
amount
}
id
title
}
}
mutation MakeBid {
bid(input: { auctionId: "1", amount: 10 }) {
amount
}
}
- I listen for changes to Auction 1 by querying the auction.
- A bid was placed on Auction 1.
- The information for Auction 1 is no longer valid.
- My query automatically refetches the latest Auction and Bid details.
Defer Directive
The @defer
directive allows you to postpone the delivery of one or more (slow) fields grouped in an inlined or spread fragment.
Slow and Fast Field Defer Example
Here, the GraphQL schema defines two queries for a "fast" and a "slow" (i.e., delayed) information.
export const schema = gql`
type Query {
"""
A field that resolves fast.
"""
fastField: String! @skipAuth
"""
A field that resolves slowly.
Maybe you want to @defer this field ;)
"""
slowField(waitFor: Int! = 5000): String @skipAuth
}
`
The Redwood services for these queries return the fastField
immediately and the showField
after some delay.
import { logger } from 'src/lib/logger'
const wait = (time: number) =>
new Promise((resolve) => setTimeout(resolve, time))
export const fastField = async () => {
return 'I am speedy'
}
export const slowField = async (_, { waitFor = 5000 }) => {
logger.debug('deferring slowField until ...')
await wait(waitFor)
logger.debug('now!')
return 'I am slow'
}
When making the query:
query SlowAndFastFieldWithDefer {
... on Query @defer {
slowField
}
fastField
}
The response returns:
{
"data": {
"fastField": "I am speedy"
}
}
and will await the deferred field to then present:
{
"data": {
"fastField": "I am speedy",
"slowField": "I am slow"
}
}
Stream Directive
The @stream
directive allows you to stream the individual items of a field of the list type as the items are available.
Alphabet Stream Example
Here, the GraphQL schema defines a query to return the letters of the alphabet:
export const schema = gql`
type Query {
alphabet: [String!]! @skipAuth
`
The service uses Repeater
to write a safe stream resolver.
AsyncGenerators as declared via the async *
keywords are prone to memory leaks and leaking timers. For real-world usage, use Repeater.
import { Repeater } from '@redwoodjs/realtime'
import { logger } from 'src/lib/logger'
export const alphabet = async () => {
return new Repeater<string>(async (push, stop) => {
const values = ['a', 'b', 'c', 'd', 'e', 'f', 'g']
const publish = () => {
const value = values.shift()
if (value) {
logger.debug({ value }, 'publishing')
push(value)
}
if (values.length === 0) {
stop()
}
}
const interval = setInterval(publish, 1000)
stop.then(() => {
logger.debug('cancel')
clearInterval(interval)
})
publish()
})
}
What does the incremental stream look like?
Since Apollo Client does not yet support the @stream
directive, you can use them in the GraphiQL Playground or see them in action via CURL.
When making the request with the @stream
directive:
curl -g -X POST \
-H "accept:multipart/mixed" \
-H "content-type: application/json" \
-d '{"query":"query StreamAlphabet { alphabet @stream }"}' \
http://localhost:8911/graphql
Here you see the initial response has []
for alphabet data.
Then on each push to the Repeater, an incremental update to the list of letters is sent.
The stream ends when hasNext
is false:
* Connected to localhost (127.0.0.1) port 8911 (#0)
> POST /graphql HTTP/1.1
> Host: localhost:8911
> User-Agent: curl/8.1.2
> accept:multipart/mixed
> content-type: application/json
> Content-Length: 53
>
< HTTP/1.1 200 OK
< connection: keep-alive
< content-type: multipart/mixed; boundary="-"
< transfer-encoding: chunked
<
---
Content-Type: application/json; charset=utf-8
Content-Length: 39
{"data":{"alphabet":[]},"hasNext":true}
---
Content-Type: application/json; charset=utf-8
Content-Length: 70
{"incremental":[{"items":["a"],"path":["alphabet",0]}],"hasNext":true}
---
Content-Type: application/json; charset=utf-8
Content-Length: 70
{"incremental":[{"items":["b"],"path":["alphabet",1]}],"hasNext":true}
---
Content-Type: application/json; charset=utf-8
Content-Length: 70
{"incremental":[{"items":["c"],"path":["alphabet",2]}],"hasNext":true}
---
Content-Type: application/json; charset=utf-8
Content-Length: 70
{"incremental":[{"items":["d"],"path":["alphabet",3]}],"hasNext":true}
---
Content-Type: application/json; charset=utf-8
Content-Length: 70
{"incremental":[{"items":["e"],"path":["alphabet",4]}],"hasNext":true}
---
Content-Type: application/json; charset=utf-8
Content-Length: 70
{"incremental":[{"items":["f"],"path":["alphabet",5]}],"hasNext":true}
---
Content-Type: application/json; charset=utf-8
Content-Length: 70
{"incremental":[{"items":["g"],"path":["alphabet",6]}],"hasNext":true}
---
...
---
Content-Type: application/json; charset=utf-8
Content-Length: 17
{"hasNext":false}
-----
How do I choose Subscriptions or Live Queries?
When deciding on how to offer realtime data updates, you’ll want to consider:
- How frequently do your users require information updates?
- Determine the value of "real-time" versus "near real-time" to your users. Do they need to know in less than 1-2 seconds, or is 10, 30, or 60 seconds acceptable for them to receive updates?
- Consider the criticality of the data update. Is it low, such as a change in shipment status, or higher, such as a change in stock price for an investment app?
- Consider the cost of maintaining connections and tracking updates across your user base. Is the infrastructure cost justifiable?
- If you don't require "real" real-time, consider polling for data updates on a reasonable interval. According to Apollo, in most cases, your client should not use subscriptions to stay up to date with your backend. Instead, you should poll intermittently with queries or re-execute queries on demand when a user performs a relevant action, such as clicking a button.
- How are you deploying? Serverless or serverful?
- Real-time options depend on your deployment method.
- If you are using a serverless architecture, your application cannot maintain a stateful connection to your users' applications. Therefore, it's not easy to "push," "publish," or "stream" data updates to the web client.
- In this case, you may need to look for third-party solutions that manage the infrastructure to maintain such stateful connections to your web client, such as Supabase Realtime, SendBird, Pusher, or consider creating your own AWS SNS-based functionality.
Showcase Demos
Please see our showcase realtime app for examples of subscriptions and live queries. It also demonstrates how you can handle streaming responses, like those used by OpenAI chat completions.
Chat Room (Subscription)
Sends a message to one of four Chat Rooms.
Each room subscribes to its new messages via the NewMessage
channel aka topic.
context.pubSub.publish('newMessage', roomId, { from, body })
Simulate
./scripts/simulate_chat.sh -h
Usage: ./scripts/simulate_chat.sh -r [roomId] -n [num_messages]
./scripts/simulate_chat.sh -h
Options:
-r roomId Specify the room ID (1-4) for sending chat messages.
-n num_messages Specify the number of chat messages to send. If not provided, the script will run with a random number of messages.
Test
/**
* To test this NewMessage subscription, run the following in one GraphQL Playground to subscribe:
*
* subscription ListenForNewMessagesInRoom {
* newMessage(roomId: "1") {
* body
* from
* }
* }
*
*
* And run the following in another GraphQL Playground to publish and send a message to the room:
*
* mutation SendMessageToRoom {
* sendMessage(input: {roomId: "1", from: "hello", body: "bob"}) {
* body
* from
* }
* }
*/
Auction Bids (Live Query)
Bid on a fancy pair of new sneaks!
When a bid is made, the auction updates via a Live Query due to the invalidation of the auction key.
const key = `Auction:${auctionId}`
context.liveQueryStore.invalidate(key)
Simulate
./scripts/simulate_bids.sh -h
Usage: ./scripts/simulate_bids.sh [options]
Options:
-a <auctionId> Specify the auction ID (1-5) for which to send bids (optional).
-n <num_bids> Specify the number of bids to send (optional).
-h, --help Display this help message.
Test
/**
* To test this live query, run the following in the GraphQL Playground:
*
* query GetCurrentAuctionBids @live {
* auction(id: "1") {
* bids {
* amount
* }
* highestBid {
* amount
* }
* id
* title
* }
* }
*
* And then make a bid with the following mutation:
*
* mutation MakeBid {
* bid(input: {auctionId: "1", amount: 10}) {
* amount
* }
* }
*/
Countdown (Streaming Subscription)
It started slowly and I thought it was my heart But then I realised that this time it was for real
Counts down from a starting values by an interval.
This example showcases how a subscription can yields its own response.
Test
/**
* To test this Countdown subscription, run the following in the GraphQL Playground:
*
* subscription CountdownFromInterval {
* countdown(from: 100, interval: 10)
* }
*/
Bedtime Story (Subscription with OpenAI Streaming)
Tell me a story about a happy, purple penguin that goes to a concert.
Showcases how to use OpenAI to stream a chat completion via a prompt that writes a bedtime story:
const PROMPT = `Write a short children's bedtime story about an Animal that is a given Color and that does a given Activity.
Give the animal a cute descriptive and memorable name.
The story should teach a lesson.
The story should be told in a quality, style and feeling of the given Adjective.
The story should be no longer than 3 paragraphs.
Format the story using Markdown.`
The story updates on each stream content delta via a newStory
subscription topic event.
context.pubSub.publish('newStory', id, story)
Movie Mashup (Live Query with OpenAI Streaming)
It's Out of Africa meets Pretty Woman.
So it's a psychic, political, thriller comedy with a heart With a heart, not unlike Ghost meets Manchurian Candidate.
-- The Player, 1992
Mashup some of your favorite movies to create something new and Netflix-worthy to watch.
Powered by OpenAI, this movie tagline and treatment updates on each stream content delta via a Live Query by invalidating the `MovieMashup key.
context.liveQueryStore.invalidate(`MovieMashup:${id}`)