Skip to content

A working compilation of RESTful design principles.

Notifications You must be signed in to change notification settings

peterboyer/restful-api-design-tips

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 

Repository files navigation

RESTful API Design Tips from Experience

A working guide of API design tips and trend evaluations.

We are all apprentices in a craft where no one ever becomes a master.

I wanted to bring together some of the many patterns I’ve come to appreciate and actively implement in present and future projects.

Contributing

If you have any comments or suggestions, please open an issue or a pull request with any changes that you would like to contribute! Thanks!

After initially publishing this article many years ago, many threads of discussion in channels such as Reddit have helped me adjust and tweak some of my explanations and stances on API design. I would like to thank all who have contributed to the discussion, and I hope this helps build this article into a more valuable resource for others!

Versioning

If you’re going to develop an API for any client service, you’re going to want to prepare yourself for eventual change. This is best realised by providing a "version-namespace" for your RESTful API.

We do this with simply adding the version as a prefix to all URLs.

GET www.myservice.com/api/v1/posts

However, through studying other API implementations, I’ve grown to like a shorter URL style offered by accessing the API as part of a subdomain, and then dropping the /api from the route; shorter and more concise is better.

GET api.myservice.com/v1/posts

Cross-Origin Resource Sharing (CORS)

It is important to consider that when placing your API into a different subdomain such as api.myservice.com it will require implementing CORS for your backend if you plan to host your frontend site at www.myservice.com and expect to use fetch requests without throwing No Access-Control-Allow-Origin header is present errors.

Methods

Use HTTP methods such as:

  • GET for fetching data.
  • POST for adding data.
  • PUT for updating data (as a whole object).
  • PATCH for updating data (with partial information for the object).
  • DELETE for deleting data.

I would like to add that I think PATCH is great way to cut down the size of requests to change parts of bigger objects, but also that it fits well with commonly implemented auto-submit/auto-save fields.

A nice example of this is with Tumblr’s "Dashboard Settings" screen, where non-critical options about the user experience of the service can be edited and saved, per item, without the need of a final form submission button. It is simply a much more organic way to interact with the user’s preference data.

img

The "Saved" tag appears and then disappears shortly after modifying the option.

Routes

When building your routes you need to think of your endpoints as groups of resources from which you may read, add, edit and delete from and these actions are encapsulated as HTTP methods.

Use Plurals

It makes semantic sense that you request many posts from /posts.

And for goodness sake don’t consider mixing /post/all with /post/:id!

Although some resource names can't be pluralised, you should try to get as close to plural as you can.

// DO: plurals are consistent and make sense
GET /v1/posts/:id/attachments/:id/comments

// DON'T: is it just one comment? is it a form? etc.
GET /v1/post/:id/attachment/:id/comment

Use Nesting for Relationship Filtering

Query strings should be used for further filtering results beyond the initial grouping of a logical set offered by a relationship.

Aim to design endpoint paths that avoid unnecessary query string parameters as they are generally harder to read and to work with when compared to paths whose structure promotes an initial relationship-based filtering and grouping of such items the deeper it goes.

This /posts/x/attachments is better than /attachments?postId=x.

And this /posts/x/attachments/y/comments is so much better than /comments?postId=x&attachmentId=y.

Use More of your "Route-Space"

You should aim to keep your API as flat as possible and not crowd your resource paths. Allow yourself to provide root-level endpoints for all of your resources for when you eventually must update/delete them. For example, the case of a post having comments, allow GET /posts/:id/comments to fetch the comments for a post based on that relationship, but also offer PATCH /comments/:id to allow editing of comments without needing to track the id for that post for every single route.

  • Longer paths for creating/fetching nested resources by relationship
  • Shorter paths for updating/deleting resources based on their id.

Use the Authorisation context to Filter

When it comes to providing an endpoint to access to all of a user's own resources (e.g. all my own posts) you may end up with many ways to serve that information; it's up to you what best suits your application.

  1. Nest a /posts relationship under /me with GET /me/posts, or
  2. Use the existing /posts endpoint but filter with query string, GET /posts?user=<id of self>, or
  3. Reuse /posts to show only your own posts, and expose public posts with GET /feed/posts.

Use a "Me" Endpoint

As hinted above, implement a GET /me endpoint to deliver basic data about the user as distinguished through the Authorisation header. This can include info about the user's permissions/scopes/groups/posts/sessions etc. that allow the client to show/hide elements and routes based on your permissions.

When it comes to providing endpoints for updating user preferences allow PATCH /me to change those intrinsic values.

Cursor-based Pagination

Pagination is necessary as your dataset grows and fetching potentially thousands of items from a collection is incredibly expensive for both the server and the client. One of the best ways to do handle pagination is by using query parameters such as from_<PROPERTY> to indicate the beginning of your result set, with an optional limit parameter that defaults to something like 20 with a maximum of 50 (depending on your use case). In your responses' meta key you can provide information such as total count, and a uniquely generated cursor used for cursor-based pagination (used by most applications, particularly Google’s Places API and Twitter).

Responses

Use Envelopes

It is good to maintain a clean response structure for your APIs in the event that you need to add metadata (e.g. pagination information) to your responses. An easy way to do this is to envelope your application response data under a data key, similar to GraphQL responses. From there you can freely add other keys, most likely an errors key in which to return any errors that are not to be conflated with correct data, and perhaps even a meta key, in which you may include pagination information like total and cursor which, once again, are not to be conflated with data items or indicative of any errors.

If you choose to build your endpoints to allow for partial submission success (in which your transaction will still operate on and return correct data, but will also accumulate errors from incorrect data to respond afterwards) you will be able to return both successful response data with data AND any problems with errors. In the case of HTTP status codes, 207 seems the most appropriate.

Separating your responses' data types into data, errors, meta, etc. helps remove validation headache when clients inspect your response, whereby you may easily check for the existence of an errors key in the response rather than using keys like code to differentiate an ITEM from an ERROR.

// DO: enveloped
{
  data: ITEM or [ ITEM, ITEM, ... ],
  meta: { total: 1000, cursor: "0794405882026854" },
  errors: [
    { code: "MY_ERROR_CODE", extensions?: { ... } },
    ERROR,
    ERROR,
    ...
  ],
}

// DON'T non-enveloped
[
  ITEM,
  ITEM,
  // ...
]

Responses and Requests

Most of the time you'll be communicating in JSON for both requests and responses.

If your API responses are in JSON, your server must include a Content-Type: application/json header expressing the MIME-type in use.

For all other response formats such as CSVs/images/etc. you must define and return those respective MIME-types in your headers.

Most modern frameworks and tools for APIs and clients-alike will automatically handle these headers for you.

Accept header

This header is set by the client making a request to the server and is used to inform the server as to what format it wants in return.

If you're only supporting JSON responses (instead of something like XML), then your API should return a 415 Unsupported Media Type HTTP status code for the resource that is trying to be accessed if the API is unable to respond in the format specified by the Accept header.

I've seen some API's that use a /json or /xml suffix as part of a URL to access a resource, however this isn't exactly RESTful, and instead the Accept header should be used.

Return the Updated Object

When updating any resource through a PUT or PATCH it’s good practice to return the updated resource in response to a successful POST , PUT , or PATCH request!

Use 204 for Deletions

Support the 204 — No Content response status code in cases where the request was successful but has no content to return. The envelope of the response, coupled with a 2XX HTTP success code is enough to indicate a successful response without arbitrary "information".

DELETE /v1/posts/:id
// response - 204
{
  "data": null
}

Use HTTP Status Codes and Error Responses

Because we are using HTTP methods, we should use HTTP status codes. Although a challenge here is to select a distinct slice of these codes, and then depend on response data to detail any response errors. Keeping a small set of codes helps you consume and handle errors consistently.

I like to use:

for Data Errors

  • 400 for when the requested information is incomplete or malformed.
  • 422 for when the requested information is okay, but invalid.
  • 404 for when everything is okay, but the resource doesn’t exist.
  • 409 for when a conflict of data exists, even with valid information.

for Auth Errors

  • 401 for when an access token isn’t provided, or is invalid.
  • 403 for when an access token is valid, but requires more privileges.

for Standard Statuses

  • 200 for when everything is okay.
  • 204 for when everything is okay, but there’s no content to return.
  • 500 for when the server throws an error, completely unexpected.

Furthermore, returning responses after these errors is also very important. I want to consider not only the presentation of the status itself, but also a reason behind it.

In the case of trying to create a new account, imagine we provide an email and password. Of course we would like to have our client app prevent any requests with an invalid email, or password that is too short, but outsiders have as much access to the API as we do from our client app when it’s live.

  • If the email field is missing, return a 400.
  • If the password field is too short, return a 422.
  • If the email field isn’t a valid email, return a 422.
  • If the email is already taken, return a 409.

"It’s much better to specify a more specific 4xx series code than just plain 400. I understand that you can put whatever you want in the response body to break down the error but codes are much easier to read at a glance." (source)

Now from these cases, two errors returned 422s regardless of their reasons being different. This is why we need an error code, and maybe even an error description. It’s important to make a distinction between code and description as I intend to have code as a machine consumable constant, and message as a human consumable string that may change.

In the case of per-field errors, the presence of the field as a key in the error is enough of a "code" to indicate that it is a target of a validation error.

Field Validation Errors

For returning those per field errors, it may be returned as:

POST /v1/register
// request
{
  "email": "end@@user.comx"
  "password": "abc"
}

// response - 422
{
  "errors": [{
    "code": "FIELDS_VALIDATION_ERROR",
    "extensions": {
      "email": "Invalid email address.",
      "password": "Password too short."
    }
  }]
}

Operational Validation Errors

And for returning operational validation errors:

POST /v1/register
// request
{
  "email": "end@user.com",
  "password": "password"
}
// response - 409
{
  "error": {
    "code": "EMAIL_ALREADY_EXISTS",
    "message": "An account already exists with this email."
  }
}

This way, your fetch logic watches out for non-200 errors, and can then straight-up check the error key from the response and then compare it to any further logic in the client app.

The message can act as a fallback human-readable error message to help understand the request when developing, and also in the case an appropriate localisation string implementation cannot be used.

Authentication

Modern stateless, RESTful APIs implement authentication with tokens most commonly provided through the Authorization header (or even an access_token query param).

Use Self-extending Session Tokens

Originally I thought that issuing JWTs for regular API requests was a great way to handle authentication—until I wanted to invalidate those tokens.

In my last revision of this post (and detailed in a separate post) I offered a way for JWTs to be reissued through an additionally stored client secret "Refresh Token" (RT) which was to be exchanged for new JWTs. However in order to expire these JWTs they each contained a reference to the issuing RT so if the RT was invalided/deleted so would the JWT. However this mechanism defeats the statelessness of the JWT itself...

My solution now is to simply use a /sessions resource endpoint to exchange login credentials for a single unique session token (using uuid4) which is hashed and stored as a database row. Just like many moderns apps, the token doesn't need to be reissued unless there is a long period of inactivity (similar to session timeout, but to the scale of weeks). After initial authentication, every future request bumps the life of the token in a self-extending manner as long as it hasn't expired.

Session Creation – Logging In

A normal login process would look like:

  1. Receive email/password combination with POST /sessions, treating sessions as just another resource.
  2. Check the email/password-hash against the database.
  3. Create a new session database row that contains a hashed uuid4() as a token.
  4. Return the non-hashed token string to the client.

Session Renewal

In this flow tokens don't need to be explicitly renewed or reissued. That's because the API extends the life of the token if its still valid every request, saving regular users from ever having a session expire for them.

Whenever a token is received by the API i.e. through an Authorization header:

  1. Receive the token i.e. from the Authorization header.
  2. Compare against the token's hash, if there is no matching session row, raise an authentication error.
  3. Check the updated_at property of the session, if now if greater than updated_at + session_life the session is considered expired, delete the session row, raise an authentication error.
  4. If it exists and is still valid from updated_at time, set the updated_at time to now() to renew the token.

Session Management

Because all sessions are tracked as database rows mapped to a user, a user can see all their active sessions similar to Facebook's account security sessions view. You can also chose to include any associated metadata you have chosen to collected when initially creating a session such as the browser's User Agent, IP address etc. And

Fetching all your sessions is as simple as:

  1. GET /sessions to return all sessions associated with your user via the Authorization header.

Session Termination – Logging Out

And because you have handles to your sessions you can terminate them to invalidate unauthorised or unwanted access to your account. And logging out would simply be terminating the client's session and purging the session from the client.

  1. Receive the token as part of a DELETE /sessions/id request.
  2. Compare against the token's hash, delete the matching session row.

Avoid Password Composition Rules

After doing a lot of research into password rules, I’ve come to agree that password rules are bullshit and are part of NIST’s "don’ts", especially considering that password composition rules help narrow down valid passwords based on their validity rules.

I’ve collated some of the best points (from the above links) for password handling:

  1. Only enforce a minimum unicode password length (min 8-10).
  2. Check against common passwords ("password12345")
  3. Check for basic entropy (don’t allow "aaaaaaaaaaaaa").
  4. Don’t use password composing rules (at least one "!@#$%&").
  5. Don’t use password hints (rhymes with "assword").
  6. Don’t use knowledge-based authentication ("security" questions).
  7. Don’t expire passwords without reason.
  8. Don’t use SMS for two-factor authentication.
  9. Use a password salt of 32-bits or more.

These "don’ts" should make password validation much easier!

Meta

Use a "Health-Check" Endpoint

Through developing with AWS, it been necessary to provide a way to output a simple response that can indicate that the API instance is alive and does not need to be restarted. It’s also useful for easily checking what version of the API is on any machine at any time, without authentication.

GET /v1
// response - 200
{
  "status": "running",
  "version": "fdb1d5e"
}

I provide status and version (which refers to the git commit ref of the API at the time it was built). It’s also worth mentioning that this value is not derived from an active .git repo being bundled with the APIs container for EC2. Instead, it is read (and stored in memory) on initialisation from a version.txt file (which is generated from the build process), and defaults to __UNKNOWN__ in case of a read error, or the file does not exist.

About

A working compilation of RESTful design principles.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published