Skip to main content
Version: 3.x

Testing

Testing. For some it's an essential part of their development workflow. For others it's something they know they should do, but for whatever reason it hasn't struck their fancy yet. For others still it's something they ignore completely, hoping the whole concept will go away. But tests are here to stay, and maybe Redwood can change some opinions about testing being awesome and fun.

Introduction to Testing

If you're already familiar with the ins and outs of testing and just want to know how to do it in Redwood, feel free to skip ahead. Or, keep reading for a refresher. In the following section, we'll build a simple test runner from scratch to help clarify the concepts of testing in our minds.

Building a Test Runner

The idea of testing is pretty simple: for each "unit" of code you write, you write additional code that exercises that unit and makes sure it works as expected. What's a "unit" of code? That's for you to decide: it could be an entire class, a single function, or even a single line! In general, the smaller the unit, the better. Your tests will stay fast and focused on just one thing, which makes them easy to update when you refactor. The important thing is that you start somewhere and codify your code's functionality in a repeatable, verifiable way.

Let's say we write a function that adds two numbers together:

const add = (a, b) => {
return a + b
}

You test this code by writing another piece of code (which usually lives in a separate file and can be run in isolation), just including the functionality from the real codebase that you need for the test to run. For our examples here we'll put the code and its test side-by-side so that everything can be run at once. Our first test will call the add() function and make sure that it does indeed add two numbers together:

const add = (a, b) => {
return a + b
}

if (add(1, 1) === 2) {
console.log('pass')
} else {
console.error('fail')
}

Pretty simple, right? The secret is that this simple check is the basis of all testing. Yes, that's it. So no matter how convoluted and theoretical the discussions on testing get, just remember that at the end of the day you're testing whether a condition is true or false.

Running a Test

You can run that code with Node or just copy/paste it into the web console of a browser. You can also run it in a dedicated web development environment like JSFiddle. Switch to the Javascript tab below to see the code:

Note that you'll see document.write() in the JSFiddle examples instead of console.log; this is just so that you can actually see something in the Result tab, which is HTML output.

You should see "pass" written to the output. To verify that our test is working as expected, try changing the + in the add() function to a - (effectively turning it into a subtract() function) and run the test again. Now you should see "fail".

Terminology

Let's get to some terminology:

  • The entire code block that checks the functionality of add() is what's considered a single test
  • The specific check that add(1, 1) === 2 is known as an assertion
  • The add() function itself is the subject of the test, or the code that is under test
  • The value you expect to get (in our example, that's the number 2) is sometimes called the expected value
  • The value you actually get (whatever the output of add(1, 1) is) is sometimes called the actual or received value
  • The file that contains the test is a test file
  • Multiple test files, all run together, is known as a test suite
  • You'll generally run your test files and suites with another piece of software. In Redwood that's Jest, and it's known as a test runner
  • The amount of code you have that is exercised by tests is referred to as coverage and is usually reported as a percentage. If every single line of code is touched as a result of running your test suite then you have 100% coverage!

This is the basic idea behind all the tests you'll write: when you add code, you'll add another piece of code that uses the first and verifies that the result is what you expect.

Tests can also help drive new development. For example, what happens to our add() function if you leave out one of the arguments? We can drive these changes by writing a test of what we want to happen, and then modify the code that's being tested (the subject) to make it satisfy the assertion(s).

Expecting Errors

So, what does happen if we leave off an argument when calling add()? Well, what do we want to happen? We'll answer that question by writing a test for what we expect. For this example let's have it throw an error. We'll write the test first that expects the error:

try {
add(1)
} catch (e) {
if (e === 'add() requires two arguments') {
console.log('pass')
} else {
console.error('fail')
}
}

This is interesting because we actually expect an error to be thrown, but we don't want that error to stop the test suite in it's tracks—we want the error to be raised, we just want to make sure it's exactly what we expect it to be! So we'll surround the code that's going to error in a try/catch block and inspect the error message. If it's what we want, then the test actually passes.

Remember: we're testing for what we want to happen. Usually you think of errors as being "bad" but in this case we want the code to throw an error, so if it does, that's actually good! Raising an error passes the test, not raising the error (or raising the wrong error) is a failure.

Run this test and what happens? (If you previously made a change to add() to see the test fail, change it back now):

Where did that come from? Well, our subject add() didn't raise any errors (Javascript doesn't care about the number of arguments passed to a function) and so it tried to add 1 to undefined, and that's Not A Number. We didn't think about that! Testing is already helping us catch edge cases.

To respond properly to this case we'll make one slight modification: add another "fail" log message if the code somehow gets past the call to add(1) without throwing an error:

try {
add(1)
console.error('fail: no error thrown')
} catch (e) {
if (e === 'add() requires two arguments') {
console.log('pass')
} else {
console.error('fail: wrong error')
}
}

We also added a little more information to the "fail" messages so we know which one we encountered. Try running that code again and you should see "fail: no error thrown" in the console.

Now we'll actually update add() to behave as we expect: by throwing an error if less than two arguments are passed.

const add = (...nums) => {
if (nums.length !== 2) {
throw 'add() requires two arguments'
}
return nums[0] + nums[1]
}

Javascript doesn't have a simple way to check how many arguments were passed to a function, so we've converted the incoming arguments to an array via spread syntax and then we check the length of that instead.

We've covered passing too few arguments, what if we pass too many? We'll leave writing that test as homework, but you should have everything you need, and you won't even need any changes to the add() function to make it work!

Our Test Runner Compared to Jest

Our tests are a little verbose (10 lines of code to test that the right number of arguments were passed). Luckily, the test runner that Redwood uses, Jest, provides a simpler syntax for the same assertions. Here's the complete test file, but using Jest's provided helpers:

describe('add()', () => {
it('adds two numbers', () => {
expect(add(1, 1)).toEqual(2)
})

it('throws an error for too few arguments', () => {
expect(() => add(1)).toThrow('add requires 2 arguments')
})
})

Jest lets us be very clear about our subject in the first argument to the describe() function, letting us know what we're testing. Note that it's just a string and doesn't have to be exactly the same as the function/class you're testing (but usually is for clarity).

Likewise, each test is given a descriptive name as the first argument to the it() functions ("it" being the subject under test). Functions like expect() and toEqual() make it clear what values we expect to receive when running the test suite. If the expectation fails, Jest will indicate that in the output letting us know the name of the test that failed and what went wrong (the expected and actual values didn't match, or an error was thrown that we didn't expect).

Jest also has a nicer output than our cobbled-together test runner using console.log:

image

Are you convinced? Let's keep going and see what Redwood brings to the table.

Redwood and Testing

Redwood relies on several packages to do the heavy lifting, but many are wrapped in Redwood's own functionality which makes them even better suited to their individual jobs:

Redwood Generators get your test suite bootstrapped. Redwood also includes Storybook, which isn't technically a test suite, but can help in other ways.

Let's explore each one and how they're integrated with Redwood.

Jest

Jest is Redwood's test runner. By default, starting Jest via yarn rw test will start a watch process that monitors your files for changes and re-runs the test(s) that are affected by that changed file (either the test itself, or the subject under test).

React Testing Library

React Testing Library is an extension of DOM Testing Library, adding functionality specifically for React. React Testing Library lets us render a single component in isolation and test that expected text is present or a certain HTML structure has been built.

Mock Service Worker

Among other things, Mock Service Worker (msw) lets you simulate the response from API calls. Where this comes into play with Redwood is how the web-side constantly calls to the api-side using GraphQL: rather than make actual GraphQL calls, which would slow down the test suite and put a bunch of unrelated code under test, Redwood uses MSW to intercept GraphQL calls and return a canned response, which you include in your test.

Storybook

Storybook itself doesn't appear to be related to testing at all—it's for building and styling components in isolation from your main application—but it can serve as a sanity check for an overlooked part of testing: the user interface. Your tests will only be as good as you write them, and testing things like the alignment of text on the page, the inclusion of images, or animation can be very difficult without investing huge amounts of time and effort. These tests are also very brittle since, depending on how they're written, they can break without any code changes at all! Imagine an integration with a CMS that allows a marketing person to make text/style changes. These changes will probably not be covered by your test suite, but could make your site unusable depending on how bad they are.

Storybook can provide a quick way to inspect all visual aspects of your site without the tried-and-true method of having a QA person log in and exercise every possible function. Unfortunately, checking those UI elements is not something that Storybook can automate for you, and so can't be part of a continuous integration system. But it makes it possible to do so, even if it currently requires a human touch.

Redwood Generators

Redwood's generators will include test files for basic functionality automatically with any Components, Pages, Cells, or Services you generate. These will test very basic functionality, but they're a solid foundation and will not automatically break as soon as you start building out custom features.

Test Commands

You can use a single command to run your entire suite :

yarn rw test

This will start Jest in "watch" mode which will continually run and monitor the file system for changes. If you change a test or the component that's being tested, Jest will re-run any associated test file. This is handy when you're spending the afternoon writing tests and always want to verify the code you're adding without swapping back and forth to a terminal and pressing Enter to run the last command again.

To start the process without watching, add the --no-watch flag:

yarn rw test --no-watch

This one is handy before committing some changes to be sure you didn't inadvertently break something you didn't expect, or before a deploy to production.

Filtering what tests to run

You can run only the web- or api-side test suites by including the side as another argument to the command:

yarn rw test web
yarn rw test api

Let's say you have a test file called CommentForm.test.js. In order to only watch and run tests in this file you can run

yarn rw test CommentForm

If you need to be more specific, you can combine side filters, with other filters

yarn rw test api Comment

which will only run test specs matching "Comment" in the API side

Testing Components

Let's start with the things you're probably most familiar with if you've done any React work (with or without Redwood): components. The simplest test for a component would be matching against the exact HTML that's rendered by React (this doesn't actually work so don't bother trying):

web/src/components/Article/Article.js
const Article = ({ article }) => {
return <article>{ article.title }</article>
}

// web/src/components/Article/Article.test.js

import { render } from '@redwoodjs/testing/web'
import Article from 'src/components/Article'

describe('Article', () => {
it('renders an article', () => {
expect(render(<Article article={ title: 'Foobar' } />))
.toEqual('<article>Foobar</article>')
})
})

This test (if it worked) would prove that you are indeed rendering an article. But it's also extremely brittle: any change to the component, even adding a className attribute for styling, will cause the test to break. That's not ideal, especially when you're just starting out building your components and will constantly be making changes as you improve them.

Why do we keep saying this test won't work?

Because as far as we can tell there's no easy way to simply render to a string. render actually returns an object that has several functions for testing different parts of the output. Those are what we'll look into in the next section.

Note that Redwood's render function is based on React Testing Library's. The only difference is that Redwood's wraps everything with mock providers for the various providers in Redwood, such as auth, the GraphQL client, the router, etc.

If you were to use React Testing Library's render function, you'd need to provide your own wrapper function. In this case you probably want to compose the mock providers from @redwoodjs/testing/web:

import { render, MockProviders } from '@redwoodjs/testing/web'

// ...

render(<Article article={ title: 'Foobar' } />, {
wrapper: ({ children }) => (
<MockProviders>
<MyCustomProvider>{children}</MyCustomProvider>
</MockProviders>
)
})

Mocking useLocation

To mock useLocation in your component tests, wrap the component with LocationProvider:

import { LocationProvider } from '@redwoodjs/router'

render(
<LocationProvider location={{ pathname: '', search: '?cancelled=true' }}>
<Component />
</LocationProvider>
)

Queries

In most cases you will want to exclude the design elements and structure of your components from your test. Then you're free to redesign the component all you want without also having to make the same changes to your test suite. Let's look at some of the functions that React Testing Library provides (they call them "queries") that let you check for parts of the rendered component, rather than a full string match.

getByText()

In our <Article> component it seems like we really just want to test that the title of the product is rendered. How and what it looks like aren't really a concern for this test. Let's update the test to just check for the presence of the title itself:

web/src/components/Article/Article.test.js
import { render, screen } from '@redwoodjs/testing/web'

describe('Article', () => {
it('renders an article', () => {
render(<Article article={ title: 'Foobar' } />)

expect(screen.getByText('Foobar')).toBeInTheDocument()
})
})

Note the additional screen import. This is a convenience helper from React Testing Library that automatically puts you in the document.body context before any of the following checks.

We can use getByText() to find text content anywhere in the rendered DOM nodes. toBeInTheDocument() is a matcher added to Jest by React Testing Library that returns true if the getByText() query finds the given text in the document.

So, the above test in plain English says "if there is any DOM node containing the text 'Foobar' anywhere in the document, return true."

queryByText()

Why not use getByText() for everything? Because it will raise an error if the text is not found in the document. That means if you want to explicitly test that some text is not present, you can't—you'll always get an error.

Consider an update to our <Article> component:

web/src/components/Article/Article.js
import { Link, routes } from '@redwoodjs/router'

const Article = ({ article, summary }) => {
return (
<article>
<h1>{article.title}</h1>
<div>
{summary ? article.body.substring(0, 100) + '...' : article.body}
{summary && <Link to={routes.article(article.id)}>Read more</Link>}
</div>
</article>
)
}

export default Article

If we're only displaying the summary of an article then we'll only show the first 100 characters with an ellipsis on the end ("...") and include a link to "Read more" to see the full article. A reasonable test for this component would be that when the summary prop is true then the "Read more" text should be present. If summary is false then it should not be present. That's where queryByText() comes in (relevant test lines are highlighted):

web/src/components/Article/Article.test.js
import { render, screen } from '@redwoodjs/testing/web'
import Article from 'src/components/Article'

describe('Article', () => {
const article = { id: 1, title: 'Foobar', body: 'Lorem ipsum...' }

it('renders the title of an article', () => {
render(<Article article={article} />)

expect(screen.getByText('Foobar')).toBeInTheDocument()
})

it('renders a summary version', () => {
render(<Article article={article} summary={true} />)

expect(screen.getByText('Read more')).toBeInTheDocument()
})

it('renders a full version', () => {
render(<Article article={article} summary={false} />)

expect(screen.queryByText('Read more')).not.toBeInTheDocument()
})
})

getByRole() / queryByRole()

getByRole() allows you to look up elements by their "role", which is an ARIA element that assists in accessibility features. Many HTML elements have a default role (including <button> and <a>) but you can also define one yourself with a role attribute on an element.

Sometimes it may not be enough to say "this text must be on the page." You may want to test that an actual link is present on the page. Maybe you have a list of users' names and each name should be a link to a detail page. We could test that like so:

it('renders a link with a name', () => {
render(<List data={[{ name: 'Rob' }, { name: 'Tom' }]} />)

expect(screen.getByRole('link', { name: 'Rob' })).toBeInTheDocument()
expect(screen.getByRole('link', { name: 'Tom' })).toBeInTheDocument()
})

getByRole() expects the role (<a> elements have a default role of link) and then an object with options, one of which is name which refers to the text content inside the element. Check out the docs for the *ByRole queries.

If we wanted to eliminate some duplication (and make it easy to expand or change the names in the future):

it('renders a link with a name', () => {
const data = [{ name: 'Rob' }, { name: 'Tom' }]

render(<List data={data} />)

data.forEach((datum) => {
expect(screen.getByRole('link', { name: datum.name })).toBeInTheDocument()
})
})

But what if we wanted to check the href of the link itself to be sure it's correct? In that case we can capture the screen.getByRole() return and run expectations on that as well (the forEach() loop has been removed here for simplicity):

import { routes } from '@redwoodjs/router'

it('renders a link with a name', () => {
render(<List data={[{ id: 1, name: 'Rob' }]} />)

const element = screen.getByRole('link', { name: data.name })
expect(element).toBeInTheDocument()
expect(element).toHaveAttribute('href', routes.user({ id: data.id }))
})

Why so many empty lines in the middle of the test?

You may have noticed a pattern of steps begin to emerge in your tests:

  1. Set variables or otherwise prepare some code
  2. render or execute the function under test
  3. expects to verify output

Most tests will contain at least the last two, but sometimes all three of these parts, and in some communities it's become standard to include a newline between each "section". Remember the acronym SEA: setup, execute, assert.

Jest Expect: Type Considerations

Redwood uses prisma as an ORM for connecting to different databases like PostgreSQL, MySQL, and many more. The database models are defined in the schema.prisma file. Prisma schema supports model field scaler types which is used to define the data types for the models properties.

Due to this, there are some exceptions that can occur while testing your API and UI components.

Floats and Decimals

Prisma recommends using Decimal instead of Float because of accuracy in precision. Float is inaccurate in the number of digits after decimal whereas Prisma returns a string for Decimal value which preserves all the digits after the decimal point.

e.g., using Float type

Expected: 1498892.0256940164
Received: 1498892.025694016

expect(result.floatingNumber).toEqual(1498892.0256940164)

e.g., using Decimal type

Expected: 7420440.088194787
Received: "7420440.088194787"

expect(result.floatingNumber).toEqual(7420440.088194787)

In the above examples, we can see expect doesn't preserve the floating numbers. Using decimals, the number is matched with the expected result.

For cases where using decimal is not optimal, see the Jest Expect documentation for other options and methods.

DateTime

Prisma returns DateTime as ISO 8601-formatted strings. So, you can convert the date to ISO String in JavaScript:

//  Output: '2021-10-15T19:40:33.000Z'
const isoString = new Date("2021-10-15T19:40:33Z").toISOString()

Other Queries/Matchers

There are several other node/text types you can query against with React Testing Library, including title, role and alt attributes, Form labels, placeholder text, and more.

If you still can't access the node or text you're looking for there's a fallback attribute you can add to any DOM element that can always be found: data-testid which you can access using getByTestId, queryByTestId and others (but it involves including that attribute in your rendered HTML always, not just when running the test suite).

You can refer to the Cheatsheet from React Testing Library with the various permutations of getBy, queryBy and siblings.

The full list of available matchers like toBeInTheDocument() and toHaveAttribute() don't seem to have nice docs on the Testing Library site, but you can find them in the README inside the main repo.

In addition to testing for static things like text and attributes, you can also use fire events and check that the DOM responds as expected.

You can read more about these in below documentations:

Mocking GraphQL Calls

If you're using GraphQL inside your components, you can mock them to return the exact response you want and then focus on the content of the component being correct based on that data. Returning to our <Article> component, let's make an update where only the id of the article is passed to the component as a prop and then the component itself is responsible for fetching the content from GraphQL:

Normally we recommend using a cell for exactly this functionality, but for the sake of completeness we're showing how to test when doing GraphQL queries the manual way!

web/src/components/Article/Article.js
import { useQuery } from '@redwoodjs/web'

const GET_ARTICLE = gql`
query getArticle($id: Int!) {
article(id: $id) {
id
title
body
}
}
`

const Article = ({ id }) => {
const { data } = useQuery(GET_ARTICLE, { variables: { id } })

if (data) {
return (
<article>
<h1>{data.article.title}</h1>
<div>{data.article.body}</div>
</article>
)
} else {
return 'Loading...'
}
}

export default Article

mockGraphQLQuery()

Redwood provides the test function mockGraphQLQuery() for providing the result of a given named GraphQL. In this case our query is named getArticle and we can mock that in our test as follows:

web/src/components/Article/Article.test.js
import { render, screen } from '@redwoodjs/testing/web'
import Article from 'src/components/Article'

describe('Article', () => {
it('renders the title of an article', async () => {
mockGraphQLQuery('getArticle', (variables) => {
return {
article: {
id: variables.id,
title: 'Foobar',
body: 'Lorem ipsum...',
}
}
})

render(<Article id={1} />)

expect(await screen.findByText('Foobar')).toBeInTheDocument()
})
})

We're using a new query here, findByText(), which allows us to find things that may not be present in the first render of the component. In our case, when the component first renders, the data hasn't loaded yet, so it will render only "Loading..." which does not include the title of our article. Without it the test would immediately fail, but findByText() is smart and waits for subsequent renders or a maximum amount of time before giving up.

Note that you need to make the test function async and put an await before the findByText() call. Read more about findBy*() queries and the higher level waitFor() utility here.

The function that's given as the second argument to mockGraphQLQuery will be sent a couple of arguments. The firstand only one we're using hereis variables which will contain the variables given to the query when useQuery was called. In this test we passed an id of 1 to the <Article> component when test rendering, so variables will contain {id: 1}. Using this variable in the callback function to mockGraphQLQuery allows us to reference those same variables in the body of our response. Here we're making sure that the returned article's id is the same as the one that was requested:

return {
article: {
id: variables.id,
title: 'Foobar',
body: 'Lorem ipsum...',
}
}

Along with variables there is a second argument: an object which you can destructure a couple of properties from. One of them is ctx which is the context around the GraphQL response. One thing you can do with ctx is simulate your GraphQL call returning an error:

mockGraphQLQuery('getArticle', (variables, { ctx }) => {
ctx.errors([{ message: 'Error' }])
})

You could then test that you show a proper error message in your component:

web/src/components/Article/Article.js
const Article = ({ id }) => {
const { data, error } = useQuery(GET_ARTICLE, {
variables: { id },
})

if (error) {
return <div>Sorry, there was an error</div>
}

if (data) {
// ...
}
}

// web/src/components/Article/Article.test.js

it('renders an error message', async () => {
mockGraphQLQuery('getArticle', (variables, { ctx }) => {
ctx.errors([{ message: 'Error' }])
})

render(<Article id={1} />)

expect(await screen.findByText('Sorry, there was an error')).toBeInTheDocument()
})

mockGraphQLMutation()

Similar to how we mocked GraphQL queries, we can mock mutations as well. Read more about GraphQL mocking in our Mocking GraphQL requests docs.

Mocking Auth

Most applications will eventually add Authentication/Authorization to the mix. How do we test that a component behaves a certain way when someone is logged in, or has a certain role?

Consider the following component (that happens to be a page) which displays a "welcome" message if the user is logged in, and a button to log in if they aren't:

web/src/pages/HomePage/HomePage.js
import { useAuth } from '@redwoodjs/auth'

const HomePage = () => {
const { isAuthenticated, currentUser, logIn } = useAuth()

return (
<>
<header>
{ isAuthenticated && <h1>Welcome back {currentUser.name}</h1> }
</header>
<main>
{ !isAuthenticated && <button onClick={logIn}>Login</button> }
</main>
</>
)
}

If we didn't do anything special, there would be no user logged in and we could only ever test the not-logged-in state:

web/src/pages/HomePage/HomePage.test.js
import { render, screen } from '@redwoodjs/testing/web'
import HomePage from './HomePage'

describe('HomePage', () => {
it('renders a login button', () => {
render(<HomePage />)

expect(screen.getByRole('button', { name: 'Login' })).toBeInTheDocument()
})
})

This test is a little more explicit in that it expects an actual <button> element to exist and that it's label (name) be "Login". Being explicit with something as important as the login button can be a good idea, especially if you want to be sure that your site is friendly to screen-readers or another assistive browsing devices.

mockCurrentUser() on the Web-side

How do we test that when a user is logged in, it outputs a message welcoming them, and that the button is not present? Similar to mockGraphQLQuery() Redwood also provides a mockCurrentUser() which tells Redwood what to return when the getCurrentUser() function of api/src/lib/auth.js is invoked:

web/src/pages/HomePage/HomePage.test.js
import { render, screen, waitFor } from '@redwoodjs/testing/web'
import HomePage from './HomePage'

describe('HomePage', () => {
it('renders a login button when logged out', () => {
render(<HomePage />)

expect(screen.getByRole('button', { name: 'Login' })).toBeInTheDocument()
})

it('does not render a login button when logged in', async () => {
mockCurrentUser({ name: 'Rob' })

render(<HomePage />)

await waitFor(() => {
expect(
screen.queryByRole('button', { name: 'Login' })
).not.toBeInTheDocument()
})
})

it('renders a welcome message when logged in', async () => {
mockCurrentUser({ name: 'Rob' })

render(<HomePage />)

expect(await screen.findByText('Welcome back Rob')).toBeInTheDocument()
})
})

Here we call mockCurrentUser() before the render() call. Right now our code only references the name of the current user, but you would want this object to include everything a real user contains, maybe an email and an array of roles.

We introduced waitFor() which waits for a render update before passing/failing the expectation. Although findByRole() will wait for an update, it will raise an error if the element is not found (similar to getByRole()). So here we had to switch to queryByRole(), but that version isn't async, so we added waitFor() to get the async behavior back.

The async behavior here is important. Even after setting the user with mockCurrentUser(), currentUser may be null during the initial render because it's being resolved. Waiting for a render update before passing/failing the exception gives the resolver a chance to execute and populate currentUser.

Figuring out which assertions need to be async and which ones don't can be frustrating, we know. If you get a failing test when using screen you'll see the output of the DOM dumped along with the failure message, which helps find what went wrong. You can see exactly what the test saw (or didn't see) in the DOM at the time of the failure.

If you see some text rendering that you're sure shouldn't be there (because maybe you have a conditional around whether or not to display it) this is a good indication that the test isn't waiting for a render update that would cause that conditional to render the opposite output. Change to a findBy* query or wrap the expect() in a waitFor() and you should be good to go!

You may have noticed above that we created two tests, one for checking the button and one for checking the "welcome" message. This is a best practice in testing: keep your tests as small as possible by only testing one "thing" in each. If you find that you're using the word "and" in the name of your test (like "does not render a login button and renders a welcome message") that's a sign that your test is doing too much.

Mocking Roles

By including a list of roles in the object returned from mockCurrentUser() you are also mocking out calls to hasRole() in your components so that they respond correctly as to whether currentUser has an expected role or not.

Given a component that does something like this:

const { currentUser, hasRole } = useAuth()

return (
{ hasRole('admin') && <button onClick={deleteUser}>Delete User</button> }
)

You can test both cases (user does and does not have the "admin" role) with two separate mocks:

mockCurrentUser({ roles: ['admin'] })
mockCurrentUser({ roles: [] })

That's it!

Handling Duplication

We had to duplicate the mockCurrentUser() call and duplication is usually another sign that things can be refactored. In Jest you can nest describe blocks and include setup that is shared by the members of that block:

describe('HomePage', () => {
describe('logged out', () => {
it('renders a login button when logged out', () => {
render(<HomePage />)

expect(screen.getByRole('button', { name: 'Login' })).toBeInTheDocument()
})
})

describe('log in', () => {
beforeEach(() => {
mockCurrentUser({ name: 'Rob' })

render(<HomePage />)
})

it('does not render a login button when logged in', async () => {
await waitFor(() => {
expect(
screen.queryByRole('button', { name: 'Login' })
).not.toBeInTheDocument()
})
})

it('renders a welcome message when logged in', async () => {
expect(await screen.findByText('Welcome back Rob')).toBeInTheDocument()
})
})
})

While the primordial developer inside of you probably breathed a sign of relief seeing this refactor, heed this warning: the more deeply nested your tests become, the harder it is to read through the file and figure out what's in scope and what's not by the time your actual test is invoked. In our test above, if you just focused on the last test, you would have no idea that currentUser is being mocked. Imagine a test file with dozens of tests and multiple levels of nested describesit becomes a chore to scroll through and mentally keep track of what variables are in scope as you look for nested beforeEach() blocks.

Some schools of thought say you should keep your test files flat (that is, no nesting) which trades ease of readability for duplication: when flat, each test is completely self contained and you know you can rely on just the code inside that test to determine what's in scope. It makes future test modifications easier because each test only relies on the code inside of itself. You may get nervous thinking about changing 10 identical instances of mockCurrentUser() but that kind of thing is exactly what your IDE is good at!

For what it's worth, your humble author endorses the flat tests style.

Testing Custom Hooks

Custom hooks are a great way to encapsulate non-presentational code. To test custom hooks, we'll use the renderHook function from @redwoodjs/testing/web.

info

Note that Redwood's renderHook function is based on React Testing Library's. The only difference is that Redwood's wraps everything with mock providers for the various providers in Redwood, such as auth, the GraphQL client, the router, etc.

If you were to use React Testing Library's renderHook function, you'd need to provide your own wrapper function. In this case you probably want to compose the mock providers from @redwoodjs/testing/web:

import { renderHook, MockProviders } from '@redwoodjs/testing/web'

// ...

renderHook(() => myCustomHook(), {
wrapper: ({ children }) => (
<MockProviders>
<MyCustomProvider>{children}</MyCustomProvider>
</MockProviders>
)
})

To use renderHook:

  1. Call your custom hook from an inline function passed to renderHook. For example:
const { result } = renderHook(() => useAccumulator(0))
  1. renderHook will return an object with the following properties:
  • result: holds the return value of the hook in its current property (so result.current). Think of result as a ref for the most recently returned value
  • rerender: a function to render the previously rendered hook with new props

Let's go through an example. Given the following custom hook:

web/src/hooks/useAccumulator/useAccumulator.js
const useAccumulator = (initialValue) => {
const [total, setTotal] = useState(initialValue)

const add = (value) => {
const newTotal = total + value
setTotal(newTotal)
return newTotal
}

return { total, add }
}

The test could look as follows:

web/src/hooks/useAccumulator/useAccumulator.test.js
import { renderHook } from '@redwoodjs/testing/web'
import { useAccumulator } from './useAccumulator'

describe('useAccumulator hook example in docs', () => {
it('has the correct initial state', () => {
const { result } = renderHook(() => useAccumulator(42))
expect(result.current.total).toBe(42)
})

it('adds a value', () => {
const { result } = renderHook(() => useAccumulator(1))
result.current.add(5)
expect(result.current.total).toBe(6)
})

it('adds multiple values', () => {
const { result } = renderHook(() => useAccumulator(0))
result.current.add(5)
result.current.add(10)
expect(result.current.total).toBe(15)
})

it('re-initializes the accumulator if passed a new initializing value', () => {
const { result, rerender } = renderHook(
(initialValue) => useAccumulator(initialValue),
{
initialProps: 0,
}
)
result.current.add(5)
rerender(99)
expect(result.current.total).toBe(99)
})
})

While renderHook lets you test a custom hook directly, there are cases where encapsulating the custom hook in a component is more robust. See https://kentcdodds.com/blog/how-to-test-custom-react-hooks.

Testing Pages & Layouts

Pages and Layouts are just regular components so all the same techniques apply!

Testing Cells

Testing Cells is very similar to testing components: something is rendered to the DOM and you generally want to make sure that certain expected elements are present.

Two situations make testing Cells unique:

  1. A single Cell can export up to four separate components
  2. There's a GraphQL query taking place

The first situation is really no different from regular component testing: you just test more than one component in your test. For example:

web/src/components/ArticleCell/ArticleCell.js
import Article from 'src/components/Article'

export const QUERY = gql`
query GetArticle($id: Int!) {
article(id: $id) {
id
title
body
}
}
`

export const Loading = () => <div>Loading...</div>

export const Empty = () => <div>Empty</div>

export const Failure = ({ error }) => <div>Error: {error.message}</div>

export const Success = ({ article }) => {
return <Article article={article} />
}

Here we're exporting four components and if you created this Cell with the Cell generator then you'll already have four tests that make sure that each component renders without errors:

web/src/components/ArticleCell/ArticleCell.test.js
import { render, screen } from '@redwoodjs/testing/web'
import { Loading, Empty, Failure, Success } from './ArticleCell'
import { standard } from './ArticleCell.mock'

describe('ArticleCell', () => {
it('renders Loading successfully', () => {
expect(() => {
render(<Loading />)
}).not.toThrow()
})

it('renders Empty successfully', async () => {
expect(() => {
render(<Empty />)
}).not.toThrow()
})

it('renders Failure successfully', async () => {
expect(() => {
render(<Failure error={new Error('Oh no')} />)
}).not.toThrow()
})

it('renders Success successfully', async () => {
expect(() => {
render(<Success article={standard().article} />)
}).not.toThrow()
})
})

You might think that "rendering without errors" is a pretty lame test, but it's actually a great start. In React something usually renders successfully or fails spectacularly, so here we're making sure that there are no obvious issues with each component.

You can expand on these tests just as you would with a regular component test: by checking that certain text in each component is present.

Cell Mocks

When the <Success> component is tested, what's this standard() function that's passed as the article prop?

If you used the Cell generator, you'll get a mocks.js file along with the cell component and the test file:

web/src/components/ArticleCell.mocks.js
export const standard = () => ({
article: {
id: 42,
}
})

Each mock will start with a standard() function which has special significance (more on that later). The return of this function is the data you want to be returned from the GraphQL QUERY defined at the top of your cell.

Something to note is that the structure of the data returned by your QUERY and the structure of the object returned by the mock is in no way required to be identical as far as Redwood is concerned. You could be querying for an article but have the mock return an animal and the test will happily pass. Redwood just intercepts the GraphQL query and returns the mock data. This is something to keep in mind if you make major changes to your QUERY—be sure to make similar changes to your returned mock data or you could get falsely passing tests!

Why not just include this data inline in the test? We're about to reveal the answer in the next section, but before we do just a little more info about working with these mocks.js file...

Once you start testing more scenarios you can add custom mocks with different names for use in your tests. For example, maybe you have a case where an article has no body, only a title, and you want to be sure that your component still renders correctly. You could create an additional mock that simulates this condition:

web/src/components/ArticleCell.mocks.js
export const standard = () => ({
article: {
id: 1,
title: 'Foobar',
body: 'Lorem ipsum...'
}
})

export const missingBody = {
article: {
id: 2,
title: 'Barbaz',
body: null
}
}

And then you just reference that new mock in your test:

web/src/components/ArticleCell/ArticleCell.test.js
import { render, screen } from '@redwoodjs/testing/web'
import { Loading, Empty, Failure, Success } from './ArticleCell'
import { standard, missingBody } from './ArticleCell.mock'

describe('ArticleCell', () => {
/// other tests...

it('Success renders successfully', async () => {
expect(() => {
render(<Success article={standard().article} />)
}).not.toThrow()
})


it('Success renders successfully without a body', async () => {
expect(() => {
render(<Success article={missingBody.article} />)
}).not.toThrow()
})
})

Note that this second mock simply returns an object instead of a function. In the simplest case all you need your mock to return is an object. But there are cases where you may want to include logic in your mock, and in these cases you'll appreciate the function container. Especially in the following scenario...

Testing Components That Include Cells

Consider the case where you have a page which renders a cell inside of it. You write a test for the page (using regular component testing techniques mentioned above). But if the page includes a cell, and a cell wants to run a GraphQL query, what happens when the page is rendered?

This is where the specially named standard() mock comes into play: the GraphQL query in the cell will be intercepted and the response will be the content of the standard() mock. This means that no matter how deeply nested your component/cell structure becomes, you can count on every cell in that stack rendering in a predictable way.

And this is where standard() being a function becomes important. The GraphQL call is intercepted behind the scenes with the same mockGraphQLQuery() function we learned about earlier. And since it's using that same function, the second argument (the function which runs to return the mocked data) receives the same arguments (variables and an object with keys like ctx).

So, all of that is to say that when standard() is called it will receive the variables and context that goes along with every GraphQL query, and you can make use of that data in the standard() mock. That means it's possible to, for example, look at the variables that were passed in and conditionally return a different object.

Perhaps you have a products page that renders either in stock or out of stock products. You could inspect the status that's passed into via variables.status and return a different inventory count depending on whether the calling code wants in-stock or out-of-stock items:

web/src/components/ProductCell/ProductCell.mock.js
export const standard = (variables) => {
return {
products: [
{
id: variables.id,
name: 'T-shirt',
inventory: variables.status === 'instock' ? 100 : 0
}
]
}
})

Assuming you had a <ProductPage> component:

web/src/pages/ProductPage/ProductPage.js
import ProductCell from 'src/components/ProductCell'

const ProductPage = ({ status }) => {
return {
<div>
<h1>{ status === 'instock' ? 'In Stock' : 'Out of Stock' }</h1>
<ProductsCell status={status} />
</div>
}
}

Which, in your page test, would let you do something like:

web/src/pages/ProductPage/ProductPage.test.js
import { render, screen } from '@redwoodjs/testing/web'
import ArticleCell from 'src/components/ArticleCell'

describe('ProductPage', () => {
it('renders in stock products', () => {
render(<ProductPage status='instock' />)

expect(screen.getByText('In Stock')).toBeInTheDocument()
})

it('renders out of stock products', async () => {
render(<ProductPage status='outofstock' />)

expect(screen.getByText('Out of Stock')).toBeInTheDocument()
})
})

Be aware that if you do this, and continue to use the standard() mock in your regular cell tests, you'll either need to start passing in variables yourself:

web/src/components/ArticleCell/ArticleCell.test.js
describe('ArticleCell', () => {
/// other tests...
test('Success renders successfully', async () => {
expect(() => {
render(<Success article={standard({ status: 'instock' }).article} />)
}).not.toThrow()
})
})

Or conditionally check that variables exists at all before basing any logic on them:

web/src/components/ArticleCell/ArticleCell.mock.js
export const standard = (variables) => {
return {
product: {
id: variables?.id || 1,
name: 'T-shirt',
inventory: variables && variables.status === 'instock' ? 100 : 0
}
}
})

Testing Forms

An alternative explanation, written in TypeScript and featuring a Storybook example, can be found on the RedwoodJS forum.

To test our forms, we can make use of of the @testing-library/user-event library which helps us approximate the the events that would actually happen in the browser if a real user were interacting with our forms. For example, calling userEvent.click(checkbox) toggles a checkbox as if a user had clicked it.

Installing @testing-library/user-event

user-event can be installed in the web side of your application by running:

yarn workspace web add -D @testing-library/user-event

Building a Form

Let's assume you've already created a component using yarn rw g component. This component is built using the @redwoodjs/forms package and provides a simple interface for using the form: we subscribe to changes via an onSubmit callback-prop.

NameForm.js
import { Form, Submit, TextField } from '@redwoodjs/forms'

const NameForm = ({ onSubmit }) => {
return (
<Form onSubmit={onSubmit}>
<TextField
name="name"
placeholder="Name"
validation={{
required: true,
}}
/>
<TextField
name="nickname"
placeholder="Nickname"
validation={{
required: false,
}}
/>
<Submit>Submit</Submit>
</Form>
)
}

export default NameForm

Testing the Form

Now, we can extend the test file which Redwood generated. We're going to want to:

1) Import waitFor from the @redwoodjs/testing/web library. 2) Add an import to @testing-library/user-event for its default. 3) Provide an onSubmit prop to our "renders successfully" test.

NameForm.test.js
import { render, screen, waitFor } from '@redwoodjs/testing/web'
import userEvent from '@testing-library/user-event'

import NameForm from './NameForm'

describe('NameForm', () => {
it('renders successfully', () => {
expect(() => {
const onSubmit = jest.fn()

render(<NameForm onSubmit={onSubmit} />)
}).not.toThrow()
})
})

Finally, we'll create three simple tests which ensure our form works as expected.

1) Does our component NOT submit when required fields are empty? 2) Does our component submit when required fields are populated? 3) Does our component submit, passing our (submit) handler the data we entered?

The important takeaways are:

  • We use await because our form's state will change multiple times; otherwise, our expect-ation would trigger prematurely.
  • We use waitFor because user-event's methods are synchronous, which contradicts the above.
    • waitFor acts as our declaration of act, required when updating the state of a React component from a test.
NameForm.test.js
import { render, screen, waitFor } from '@redwoodjs/testing/web'
import userEvent from '@testing-library/user-event'

import NameForm from './NameForm'

describe('NameForm', () => {

it('does not submit when required fields are empty', async () => {
const onSubmit = jest.fn()

render(<NameForm onSubmit={onSubmit} />)

const submitButton = screen.getByText('Submit')

await waitFor(() => userEvent.click(submitButton))

expect(onSubmit).not.toHaveBeenCalled()
})

it('submits when required fields are entered', async () => {
const name = 'My Name'
const nickname = ''

const onSubmit = jest.fn()

render(<NameForm onSubmit={onSubmit} />)

const nameField = screen.getByPlaceholderText('Name')
const submitButton = screen.getByText('Submit')

await waitFor(() => userEvent.type(nameField, name))
await waitFor(() => userEvent.click(submitButton))

expect(onSubmit).toHaveBeenCalledTimes(1)
expect(onSubmit).toHaveBeenCalled()
expect(onSubmit).toHaveBeenCalledWith(
{ name, nickname },
expect.objectContaining({
_reactName: 'onSubmit',
type: 'submit',
})
)
})

it('submits with the expected, entered data', async () => {
const name = 'My Name'
const nickname = 'My Nickname'

const onSubmit = jest.fn()

render(<NameForm onSubmit={onSubmit} />)

const nameField = screen.getByPlaceholderText('Name')
const nicknameField = screen.getByPlaceholderText('Nickname')
const submitButton = screen.getByText('Submit')

await waitFor(() => userEvent.type(nameField, name))
await waitFor(() => userEvent.type(nicknameField, nickname))
await waitFor(() => userEvent.click(submitButton))

expect(onSubmit).toHaveBeenCalledTimes(1)
expect(onSubmit).toHaveBeenCalled()
expect(onSubmit).toHaveBeenCalledWith(
{ name, nickname },
expect.objectContaining({
_reactName: 'onSubmit',
type: 'submit',
})
)
})

// })

Testing Services

Until now we've only tested things on the web-side of our app. When we test the api-side that means testing our Services.

In some ways testing a Service feels more "concrete" than testing components—Services deal with hard data coming out of a database or third party API, while components deal with messy things like language, layout, and even design elements.

Services will usually contain most of your business logic which is important to verify for correctness—crediting or debiting the wrong account number on the Services side could put a swift end to your business!

The Test Database

To simplify Service testing, rather than mess with your development database, Redwood creates a test database that it executes queries against. By default this database will be located at the location defined by a TEST_DATABASE_URL environment variable and will fall back to .redwood/test.db if that var does not exist.

If you're using Postgres or MySQL locally you'll want to set that env var to your connection string for a test database in those services.

info

Does anyone else find it confusing that the software itself is called a "database", but the container that actually holds your data is also called a "database," and you can have multiple databases (containers) within one instance of a database (software)?

When you start your test suite you may notice some output from Prisma talking about migrating the database. Redwood will automatically run yarn rw prisma db push against your test database to make sure it's up-to-date.

What if I have custom migration SQL?

The prisma db push command only restores a snapshot of the current database schema (so that it runs as fast as possible). It does not actually run migrations in sequence. This can cause a problem if you have certain database configuration that must occur as a result of the SQL statements inside the migration files.

In order to preserve those statements in your test database, you can set an additional ENV var which will use the command yarn rw prisma migrate reset instead. This will run each migration in sequence against your test database. The tradeoff is that starting your test suite will take a little longer depending on how many migrations you have:

/.env
TEST_DATABASE_STRATEGY=reset

Set the variable to push, or remove it completely, and it will use the default behavior of running yarn rw prisma db push.

Writing Service Tests

A Service test can be just as simple as a component test:

api/src/services/users/users.js
export const createUser = ({ input }) => {
return db.user.create({ data: input })
}

// api/src/services/users/users.test.js
import { createUser } from './users'

describe('users service', () => {
it('creates a user', async () => {
const record = await createUser({ name: 'David' })

expect(record.id).not.toBeNull()
expect(record.name).toEqual('David')
})
})

This test creates a user and then verifies that it now has an id and that the name is what we sent in as the input. Note the use of async/await: although the service itself doesn't use async/await, when the service is invoked as a GraphQL resolver, the GraphQL provider sees that it's a Promise and waits for it to resolve before returning the response. We don't have that middleman here in the test suite so we need to await manually.

Did a user really get created somewhere? Yes: in the test database!

In theory, it would be possible to mock out the calls to db to avoid talking to the database completely, but we felt that the juice wouldn't be worth the squeeze—you will end up mocking tons of functionality that is also under active development (Prisma) and you'd constantly be chasing your tail trying to keep up. So we give devs a real database to access and remove a whole class of frustrating bugs and false test passes/failures because of out-of-date mocks.

Database Seeding

What about testing code that retrieves a record from the database? Easy, just pre-seed the data into the database first, then test the retrieval. Seeding refers to setting some data in the database that some other code requires to be present to get its job done. In a production deployment this could be a list of pre-set tags that users can apply to forum posts. In our tests it refers to data that needs to be present for our actual test to use.

In the following code, the "David" user is the seed. What we're actually testing is the users() and user() functions. We verify that the data returned by them matches the structure and content of the seed:

it('retrieves all users', async () => {
const data = await createUser({ name: 'David' })

const list = await users({ id: data.id })

expect(list.length).toEqual(1)
})

it('retrieves a single user', async () => {
const data = await createUser({ name: 'David' })

const record = await user({ id: data.id })

expect(record.id).toEqual(data.id)
expect(record.name).toEqual(data.name)
})

Notice that the string "David" only appears once (in the seed) and the expectations are comparing against values in data, not the raw strings again. This is a best practice and makes it easy to update your test data in one place and have the expectations continue to pass without edits.

Did your spidey sense tingle when you saw that exact same seed duplicated in each test? We probably have other tests that check that a user is editable and deletable, both of which would require the same seed again! Even more tingles! When there's obvious duplication like this you should know by now that Redwood is going to try and remove it.

Scenarios

Redwood created the concept of "scenarios" to cover this common case. A scenario is a set of seed data that you can count on existing at the start of your test and removed again at the end. This means that each test lives in isolation, starts with the exact same database state as every other one, and any changes you make are only around for the length of that one test, they won't cause side-effects in any other.

When you use any of the generators that create a service (scaffold, sdl or service) you'll get a scenarios.js file alongside the service and test files:

export const standard = defineScenario({
user: {
one: {
data: {
name: 'String',
},
},
two: {
data: {
name: 'String',
},
}
},
})

This scenario creates two user records. The generator can't determine the intent of your fields, it can only tell the datatypes, so strings get prefilled with just 'String'. What's up with the one and two keys? Those are friendly names you can use to reference your scenario data in your test.

The data key is one of Prisma's create options. It's the same as in your Services—everything in the one and two keys actually just gets passed to Prisma's create. You can even create relationships if you want.

Let's look at a better example. We'll update the scenario with some additional data and give them a more distinctive name:

export const standard = defineScenario({
user: {
anthony: {
data : {
name: 'Anthony Campolo',
email: 'anthony@redwoodjs.com'
},
},
dom: {
data: {
name: 'Dom Saadi',
email: 'dom@redwoodjs.com'
},
}
},
})

Note that even though we are creating two users we don't use array syntax and instead just pass several objects. Why will become clear in a moment.

Now in our test we replace the it() function with scenario():

scenario('retrieves all users', async (scenario) => {
const list = await users()

expect(list.length).toEqual(Object.keys(scenario.user).length)
})

scenario('retrieves a single user', async (scenario) => {
const record = await user({ id: scenario.user.dom.id })

expect(record.id).toEqual(scenario.user.dom.id)
})

The scenario argument passed to the function contains the scenario data after being inserted into the database which means it now contains the real id that the database assigned the record. Any other fields that contain a database default value will be populated, included DateTime fields like createdAt. We can reference individual model records by name, like scenario.user.dom. This is why scenario records aren't created with array syntax: otherwise we'd be referencing them with syntax like scenario.user[1]. Yuck.

Named Scenarios

You may have noticed that the scenario we used was once again named standard. This means it's the "default" scenario if you don't specify a different name. This implies that you can create more than one scenario and somehow use it in your tests. And you can:

export const standard = defineScenario({
user: {
anthony: {
data : {
name: 'Anthony Campolo',
email: 'anthony@redwoodjs.com'
},
},
dom: {
data: {
name: 'Dom Saadi',
email: 'dom@redwoodjs.com'
},
}
},
})

export const incomplete = defineScenario({
user: {
david: {
data: {
name: 'David Thyresson',
email: 'dt@redwoodjs.com'
},
},
forrest: {
data: {
name: '',
email: 'forrest@redwoodjs.com'
},
}
}
})
scenario('incomplete', 'retrieves only incomplete users', async (scenario) => {
const list = await users({ complete: false })
expect(list).toMatchObject([scenario.user.forrest])
})

The name of the scenario you want to use is passed as the first argument to scenario() and now those will be the only records present in the database at the time the test is run. Assume that the users() function contains some logic to determine whether a user record is "complete" or not. If you pass { complete: false } then it should return only those that it determines are not complete, which in this case includes users who have not entered their name yet. We seed the database with the scenario where one user is complete and one is not, then check that the return of users() only contains the user without the name.

Multiple Models

You're not limited to only creating a single model type in your scenario, you can populate every table in the database if you want.

export const standard = defineScenario({
product: {
shirt: {
data: {
name: 'T-shirt',
inventory: 5
},
}
},
order: {
first: {
data: {
poNumber: 'ABC12345'
},
}
},
paymentMethod: {
credit: {
data: {
type: 'Visa',
last4: 1234
},
}
}
})

And you reference all of these on your scenario object as you would expect

scenario.product.shirt
scenario.order.first
scenario.paymentMethod.credit

Relationships

What if your models have relationships to each other? For example, a blog Comment has a parent Post. Scenarios are passed off to Prisma's create function, which includes the ability to create nested relationship records simultaneously:

export const standard = defineScenario({
comment: {
first: {
data: {
name: 'Tobbe',
body: 'But it uses some letters twice'
post: {
create: {
title: 'Every Letter',
body: 'The quick brown fox jumped over the lazy dog.'
}
}
},
}
}
})

Now you'll have both the comment and the post it's associated to in the database and available to your tests. For example, to test that you are able to create a second comment on this post:

scenario('creates a second comment', async (scenario) => {
const comment = await createComment({
input: {
name: 'Billy Bob',
body: "A tree's bark is worse than its bite",
postId: scenario.comment.jane.postId,
},
})

const list = await comments({ postId: scenario.comment.jane.postId })

expect(list.length).toEqual(Object.keys(scenario.comment).length + 1)
})

postId is created by Prisma after creating the nested post model and associating it back to the comment.

Why check against Object.keys(scenario.comment).length + 1 and not just 2? Because if we ever update the scenario to add more records (maybe to support another test) this test will keep working because it only assumes what it itself did: add one comment to existing count of comments in the scenario.

You can also include the post object (or select specific fields from it):

export const standard = defineScenario({
comment: {
first: {
data: {
name: 'Rob',
body: 'Something really interesting'
post: {
create: {
title: 'Brand new post',
body: 'Introducing dbAuth'
}
}
},
include: {
post: true
}
}
}
})

Then you’ll have both the comment and its post:

scenario('retrieves a comment with post', async (scenario) => {
const comment = await commentWithPost({ id: scenario.comment.first.id })

expect(comment.post.title).toEqual(scenario.comment.first.post.title)
})

Relationships with Existing Records

If your models have relationships and you need to connect new records to existing ones, using the object syntax just isn't going to cut it.

Consider a Comment: it has a parent Post, and both of them have an Author. Using the object syntax, there's no way of accessing the authorId of the Author we just created. We could potentially hardcode it, but that's bad practice.

export const standard = defineScenario({
post: {
first: {
data: {
name: 'First Post',
author: { create: { name: 'Kris' }},
comments: {
create: [
{
name: 'First Comment',
body: 'String',
authorId: // Here we want a different author...
},
{
name: 'First Comment Response',
body: 'String',
authorId: // But here we want the same author as the post...
},
],
},
}
}),
},
})

When you run into this, you can access an existing scenario record using the distinctive name key as a function that returns an object:

export const standard = defineScenario({
author: {
kris: {
data: { name: 'Kris' }
}
rob: {
data: { name: 'rob' }
}
},
post: {
first: (scenario) => ({
data: {
name: 'First Post',
authorId: scenario.author.kris.id,
comments: {
create: [
{
name: 'First Comment',
body: 'String',
authorId: scenario.author.rob.id,
},
{
name: 'First Comment Response',
body: 'String',
authorId: scenario.author.kris.id,
},
],
},
}
}),
},
})

Since ES2015, object property keys are in ascending order of creation. This means that a key in defineScenario has access to key(s) created before it. We can leverage this like so:

export const standard = defineScenario({
user: {
kris: {
data: { name: 'Kris' }
}
},
post: {
first: (scenario) => ({
// Here you have access to the user above via `scenario.user`
}),
},
comment: {
first: (scenario) => ({
// Here you have access to both `scenario.user` and `scenario.post`
})
}
})
tip

Looking for info on how TypeScript works with Scenarios? Check out the Utility Types doc

Which Scenarios Are Seeded?

Only the scenarios named for your test are included at the time the test is run. This means that if you have:

  • posts.test.js
  • posts.scenarios.js
  • comments.test.js
  • comments.scenarios.js

Only the posts scenarios will be present in the database when running the posts.test.js and only comments scenarios will be present when running comments.test.js. And within those scenarios, only the standard scenario will be loaded for each test unless you specify a differently named scenario to use instead.

During the run of any single test, there is only ever one scenario's worth of data present in the database: users.standard or users.incomplete.

mockCurrentUser() on the API-side

Just like when testing the web-side, we can use mockCurrentUser() to mock out the user that's currently logged in (or not) on the api-side.

Let's say our blog, when commenting, would attach a comment to a user record if that user was logged in while commenting. Otherwise the comment would be anonymous:

api/src/services/comments/comments.js
export const createComment = ({ input }) => {
if (context.currentUser) {
return db.comment.create({ data: { userId: context.currentUser.id, ...input }})
} else {
return db.comment.create({ data: input })
}
}

We could include a couple of tests that verify this functionality like so:

api/src/services/comments/comments.test.js
scenario('attaches a comment to a logged in user', async (scenario) => {
mockCurrentUser({ id: 123, name: 'Rob' })

const comment = await createComment({
input: {
body: "It is the nature of all greatness not to be exact.",
postId: scenario.comment.jane.postId,
},
})

expect(comment.userId).toEqual(123)
})

scenario('creates anonymous comment if logged out', async (scenario) => {
// currentUser will return `null` by default in tests, but it's
// always nice to be explicit in tests that are testing specific
// behavior (logged in or not)—future devs may not go in with the
// same knowledge/assumptions as us!
mockCurrentUser(null)

const comment = await createComment({
input: {
body: "When we build, let us think that we build for ever.",
postId: scenario.comment.jane.postId,
},
})

expect(comment.userId).toEqual(null)
})

Testing Functions

Testing serverless functions and webhooks can be difficult and time-consuming because you have to construct the event and context information that the function handler needs.

Webhook testing is even more complex because you might need to open a http tunnel to a running dev server to accept an incoming request, then you'll have to sign the webhook payload so that the request is trusted, and then you might even trigger events from your third-party service ... all manually. Every. Time.

Luckily, RedwoodJS has several api testing utilities to make testing functions and webhooks a breeze -- and without having to run a dev server.

Want to learn to How to Test Serverless Functions and Webhooks?

We have an entire testing section in the Serverless Functions documentation that will walk your through an example of a function and a webhook.

Testing GraphQL Directives

Please refer to the Directives documentation for details on how to write Redwood Validator or Transformer Directives tests.

Testing Caching

If you're using Redwood's caching, we provide a handful of utilities and patterns to help you test this too!

Let's say you have a service where you cache the result of products, and individual products:

export const listProducts: QueryResolvers['listProducts'] = () => {
return cacheFindMany('products-list', db.product, {
expires: 3600,
})
}

export const product: QueryResolvers['product'] = async ({ id }) => {
return cache(
`cached-product-${id}`,
() =>
db.product.findUnique({
where: { id },
}),
{ expires: 3600 }
)
}

With this code, we'll be caching an array of products (from the find many), and individual products that get queried too.

tip

It's important to note that when you write scenario or unit tests, it will use the InMemoryClient.

The InMemoryClient has a few extra features to help with testing.

  1. Allows you to call cacheClient.clear() so each of your tests have a fresh cache state
  2. Allows you to get all its contents (without cache-keys) with the cacheClient.contents getter

There's a few different things you may want to test, but let's start with the basics.

In your test let's import your cache client and clear after each test:

import type { InMemoryClient } from '@redwoodjs/api/cache'
import { client } from 'src/lib/cache'

// For TypeScript users
const testCacheClient = client as InMemoryClient

describe('products', () => {
afterEach(() => {
testCacheClient.clear()
})
//....
})

The toHaveCached matcher

We have a custom Jest matcher included in Redwood to make things a little easier. To use it simply add an import to the top of your test file:

import '@redwoodjs/testing/cache'
// ^^ make `.toHaveCached` available

The toHaveCached matcher can take three forms:

expect(testCacheClient)

  1. .toHaveCached(expectedData) - check for an exact match of the data, regardless of the key
  2. .toHaveCached('expected-key', expectedData) - check that the data is cached in the key you supply
  3. .toHaveCached(/key-regex.*/, expectedData) - check that data is cached in a key that matches the regex supplied

Let's see these in action now:

scenario('returns a single product', async (scenario: StandardScenario) => {
await product({ id: scenario.product.three.id })

// Pattern 1: Only check that the data is present in the cache
expect(testCacheClient).toHaveCached(scenario.product.three)

// Pattern 2: Check that data is cached, at a specific key
expect(testCacheClient).toHaveCached(
`cached-product-${scenario.product.three.id}`,
scenario.product.three
)

// Pattern 3: Check that data is cached, in a key matching the regex
expect(testCacheClient).toHaveCached(
/cached-.*/,
scenario.product.three
)
Serialized Objects in Cache

Remember that the cache only ever contains serialized objects. So if you passed an object like this:

{
id: 5,
published: new Date('12/10/1995')
}

The published key will be serialized and stored as a string. To make testing easier for you, we serialize the object you are passing when you use the toHaveCached matcher, before we compare it against the value in the cache.

Partial Matching

It can be a little tedious to check that every key in the object you are looking for matches. This is especially true if you have autogenerated values such as updatedAt and cuid IDs.

To help with this, we've provided a helper for partial matching!

import { partialMatch } from '@redwoodjs/testing/cache'

scenario('returns all products', async (scenario: StandardScenario) => {
await products()

// Partial match using the toHaveCached, if you supply a key
expect(testCacheClient).toHaveCached(
/cached-products.*/,
partialMatch([{ name: 'LS50', brand: 'KEF' }])
)

// Or you can use the .contents getter
expect(testCacheClient.contents).toContainEqual(
// check that an array contains an object matching
partialMatch([{ name: 'LS50', brand: 'KEF' }])
)
}

scenario('finds a single product', () = {
await product({id: 5})

// You can also check for a partial match of an object
expect(testCacheClient).toHaveCached(
/cached-.*/,
partialMatch({
name: 'LS50',
brand: 'KEF'
})
)
})

Partial match is just syntactic sugar—underneath it uses Jest's expect.objectContaining and expect.arrayContaining.

The partialMatch helper takes two forms of arguments:

  • If you supply an object, you are expecting a partial match of that object
  • If you supply an array of objects, you are expecting an array containing a partial match of each of the objects
tip

Note that you cannot use partialMatch with toHaveCached without supplying a key!

// 🛑 Will never pass!
expect(testCacheClient).toHaveCached(partialMatch({name: 'LS50'}))

For partial matches, you either have to supply a key to toHaveCached or use the cacheClient.contents helper.

Strict Matching

If you'd like stricter checking (i.e. you do not want helpers to automatically serialize/deserialize your expected value), you can use the .contents getter in test cache client. Note that the .contents helper will still de-serialize the values in your cache (to make it easier to compare), just not the expected value.

For example:


const expectedValue = {
// Note that this is a date 👇
publishDate: new Date('12/10/1988'),
title: 'A book from the eighties',
id: 1988
}

// ✅ will pass, because we will serialize the publishedDate for you
expect(testCacheClient).toHaveCached(expectedValue)


// 🛑 won't pass, because publishDate in cache is a string, but you supplied a Date object
expect(testCacheClient.contents).toContainEqual(expectedValue)

// ✅ will pass, because you serialized the date
expect(testCacheClient.contents).toContainEqual({
...expectedValue,
publishDate: expectedValue.publishDate.toISOString()
})

// And if you wanted to view the raw contents of the cache
console.log(testCacheClient.storage)

This is mainly helpful when you are testing for a very specific value, or have edgecases in how the serialization/deserialization works in the cache.

Wrapping Up

So that's the world of testing according to Redwood. Did we miss anything? Can we make it even more awesome? Stop by the community and ask questions, or if you've thought of a way to make this doc even better then open a PR.

Now go out and create (and test!) something amazing!