The content is indexed in a general search engine and findable when you create searches with the product name and some key tasks.
Assessment:
Score: 0-5
The doc site has a site-specific search that lets users search within the documentation site itself.
Assessment:
Score: 0-5
If you have lots of documentation relevant to users, the documentation is organized in a developer portal of some kind.
Assessment:
Score: 0-5
If there's a UI (such as with an administrative console), links within the UI connect users to the relevant documentation.
Assessment:
Score: 0-5
The content doesn't have an endless FAQ with information that should have been integrated into more logical places in the document.
Assessment:
Score: 0-5
If content has multiple versions, the versions are called out visibly in the topic and might have a selector or link allows users to navigate to the other versions.
Assessment:
Score: 0-5
Release notes track changes to the product and documentation.
Assessment:
Score: 0-5
There’s an easy path for users to find the top 5 most-visited pages on the site. This requires you to look at metrics to determine these pages, and then assess the flow to those pages.
Assessment:
Score: 0-5
Content has been reviewed by a subject matter expert within the past year. Ideally, each topic should include metadata such as the last-reviewed timestamp, last author, and the group that owns the content.
Assessment:
Score: 0-5
The documentation provides a visible timestamp of the last time it was edited so that users can gauge how current the documentation is.
Assessment:
Score: 0-5
Links point to correct pages or are appropriately handled by redirects to equivalent pages.
Assessment:
Score: 0-5
The documentation doesn't just provide reference information (e.g., auto-generated API documentation) but also explains how to use the API with tutorials guiding users through common use cases and journeys. The content should address the *most common* use cases intended for the product.
Assessment:
Score: 0-5
Code samples](docapis_codesamples_bestpractices.html) showing sample ways to use the API (or similar tools) are provided. Ideally, the code samples are available in the user's target language. This might mean providing multiple code samples.
Assessment:
Score: 0-5
Options for contact or support are provided, even if the support merely involves posting to a peer-monitored forum.
Assessment:
Score: 0-5
The docs allow users to interact with the functions, API calls, or other commands often using their custom data. Typically, with REST APIs, an API explorer allows users to try out requests using their own data. Other sites might use Jupyter Notebooks to run code that can be customized. Seeing responses with your own data can make the API seem more relevant.
Assessment:
Score: 0-5
A feedback option is included in the docs so readers can relay feedback about their experience. When users provide feedback, they get the sense that someone is listening. This helps users feel more relevant.
Assessment:
Score: 0-5
The overview explains the big picture and describes the problem that the tool or service addresses. Common who/what/where/why questions are answered here.
Assessment:
Score: 0-5
The overview provides a high-level description of the main use cases or business objectives of the product. This allows users to get a sense of what the API is all about.
Assessment:
Score: 0-5
The overview has a diagram of the main components and how they interact. This provides users with a glimpse of the whole.
Assessment:
Score: 0-5
If there's an SDK or developer kit that users can download, the contents of this download are described. This is similar to product instructions that start by identifying all parts that should have arrived in a package.
Assessment:
Score: 0-5
For larger systems that might have multiple subsystems (e.g., groups of APIs for different scenarios), these subsystems have their own landing pages that resemble the higher-level overview (with use cases, diagrams, getting started links) but scoped to that specific subsystem.
Assessment:
Score: 0-5
Details about how to get access, permissions, and authorization to use the API are provided. For example, this topic might cover how to authorize an API call with API keys.
Assessment:
Score: 0-5
A getting started tutorial is provided for users to get started in an end-to-end way with the product, producing a sample output that builds their confidence. This topic might provide info on how to sign up, register, get API keys or permissions, and start using the API. (This topic might link to the authorization topic but is more comprehensive in scope. The purpose of this topic is frictionless onboarding.)
Assessment:
Score: 0-5
A sandbox environment is available for users to try out the tool/system/API on their own and get a better understanding of how it works.
Assessment:
Score: 0-5
If the product has a code repository, the repository contains a README file that provides information similar to the overview (but abbreviated), including information on how to contribute if applicable. (The README might simply link to the documentation overview but also include unique sections such as how to contribute in the repo.)
Assessment:
Score: 0-5
Each task has a "Prerequisites" section that explains knowledge requirements, tool requirements, essential concepts, etc., necessary for completing the task.
Assessment:
Score: 0-5
Tasks are articulated through a series of specific, usually numbered steps or sections that reduce the complexity of the task. More complex tasks might require more granular and numerous steps or sections so as to reduce the complexity. (This is the primary way that technical writers simplify complexity.)
Assessment:
Score: 0-5
In more robust docs, a sample app might serve as a reference implementation for users.
Assessment:
Score: 0-5
Code samples are appropriately commented, either with inline comments, comments that preface or follow the sample, or both. Long code samples with no explanation aren't just dropped in.
Assessment:
Score: 0-5
For native library APIs (or other API types), reference docs (auto-generated from source code comments) are available. This might mean nativelibraryapis_exploring_javadoc_output.html, Doxygen, OpenAPI outputs like Swagger or other reference docs specific to the library. The reference docs should be populated and generally follow tagging standards.
Assessment:
Score: 0-5
Reference material such as APIs follow a common structure within each topic, mostly following a request-response type structure. Typical sections include descriptions, parameters, sample requests or usage, and sample responses.
Assessment:
Score: 0-5
Error messages that users can encounter are documented and discoverable through search.
Assessment:
Score: 0-5
Parameter documentation typically includes a description, data type, min/max values, sample values, and optional/required usage.
Assessment:
Score: 0-5
The response documentation for REST APIs provides both a sample response and schema. The response provides an example of what might be returned, while the schema defines all possible elements that might be returned and describes attributes such as data types and whether the elements are required or optional in the response.
Assessment:
Score: 0-5
The documentation has a troubleshooting section (either standalone or included within the section/topic it relates to) that provides information on how to solve common problems. The troubleshooting information indicates where things might go wrong and how to fix them. In other words, the documentation doesn't just describe the happy path.
Assessment:
Score: 0-5
If there are limitations and thresholds (for example, max number of API calls, max file sizes, max string lengths, max/min parameter values, max server load), these limitations are called out in the docs.
Assessment:
Score: 0-5
If a feature is available only in certain contexts (locales, languages, platforms, roles, versions), that information is noted clearly in the feature. For example, an API that is only available for enterprise versions might have a label that says "Enterprise Version Only," or if only available for a particular platform, might say "Linux Only" or the equivalent.
Assessment:
Score: 0-5
The sidebar navigation lets users take in a sense of the whole while also allowing users to expand more details as desired. The sidebar isn't a massive list of seemingly endless scrolling and expansion + expansion + expansion but divides up doc sets into logical groups, like chapters in a book. For systems with large numbers of topics, progressive disclose techniques might be implemented across primary, secondary, and tertiary levels of information.
Assessment:
Score: 0-5
As the user navigates each topic, the sidebar navigation makes it clear where the user is in the navigation (for example, the topic highlights clearly and the navigation sticks open at that level). Breadcrumbs might also help establish site context.
Assessment:
Score: 0-5
When a user clicks topics in the navigation, the UI doesn't shift context in jarring ways, such as unexpectedly taking the user to another doc set or changing stable navigation areas like the sidebar and header (which are usually consistent for every page). This jarring navigation often happens when sidebar entries point to topics in other doc sites. If this is the case, the external links have an icon indicating the link takes them to another site.
Assessment:
Score: 0-5
Common topics have similar names across doc sets in the developer portal. For example, the Overview, Getting Started, Troubleshooting, Glossary, Release Notes, and Reference are named consistently to help users understand how to navigate the site. One doc set shouldn't call a topic "Latest updates" and "First steps" while another uses "What's new" and "Quickstart."
Assessment:
Score: 0-5
The subheadings are descriptive enough to allow users to get a gist of the topic by merely reading the subheadings. The subheadings also follow a parallel structure so they can be more easily scanned and read.
Assessment:
Score: 0-5
Each topic usually has a short summary below the title that encapsulates the main purpose of the topic. This helps users get a sense of the topic at a glance.
Assessment:
Score: 0-5
Unfamiliar words and jargon are defined in a glossary. At times, the glossary terms are linked to their glossary definitions.
Assessment:
Score: 0-5
The formatting in code samples follows standard white spacing, line breaks, and other syntax for the language. Code syntax highlighting appropriate to the language has been applied to increase the code's readability.
Assessment:
Score: 0-5
The content can be read on a mobile device (e.g., iPhone) in a usable way. For example, the responsive view allows users to navigate the sidebar links and view code samples.
Assessment:
Score: 0-5