12 breaking releases

new 0.14.0 Dec 2, 2024
0.13.0 Nov 13, 2024
0.12.0 Sep 26, 2024
0.10.1 May 15, 2024
0.1.0 Jun 25, 2020

#406 in HTTP server

Download history 1058/week @ 2024-08-12 1562/week @ 2024-08-19 3584/week @ 2024-08-26 3058/week @ 2024-09-02 2791/week @ 2024-09-09 2880/week @ 2024-09-16 5731/week @ 2024-09-23 6570/week @ 2024-09-30 4068/week @ 2024-10-07 4564/week @ 2024-10-14 2923/week @ 2024-10-21 3793/week @ 2024-10-28 3355/week @ 2024-11-04 4926/week @ 2024-11-11 5523/week @ 2024-11-18 3270/week @ 2024-11-25

17,179 downloads per month
Used in 7 crates

Apache-2.0

570KB
11K SLoC

Dropshot is a general-purpose--but opinionated--crate for exposing REST APIs from a Rust program. It is meant to be simple and lightweight. It includes first-class support for OpenAPI in the form of precise specs generated directly from code. It supports primitives for consistent pagination including extensions an OpenAPI extension denoting it in the spec.

For information about configuration, design, and contributing, see the GitHub repo.

For information on the use the API see the documentation.


lib.rs:

Dropshot is a general-purpose crate for exposing REST APIs from a Rust program. Planned highlights include:

  • Suitability for production use on a largely untrusted network. Dropshot-based systems should be high-performing, reliable, debuggable, and secure against basic denial of service attacks (intentional or otherwise).

  • First-class OpenAPI support, in the form of precise OpenAPI specs generated directly from code. This works because the functions that serve HTTP resources consume arguments and return values of specific types from which a schema can be statically generated.

  • Ease of integrating into a diverse team. An important use case for Dropshot consumers is to have a team of engineers where individuals might add a few endpoints at a time to a complex server, and it should be relatively easy to do this. Part of this means an emphasis on the principle of least surprise: like Rust itself, we may choose abstractions that require more time to learn up front in order to make it harder to accidentally build systems that will not perform, will crash in corner cases, etc.

By "REST API", we primarily mean an API built atop existing HTTP primitives, organized into hierarchical resources, and providing consistent, idempotent mechanisms to create, update, list, and delete those resources. "REST" can mean a range of things depending on who you talk to, and some people are dogmatic about what is or isn't RESTy. We find such dogma not only unhelpful, but poorly defined. (Consider such a simple case as trying to update a resource in a REST API. Popular APIs sometimes use PUT, PATCH, or POST for the verb; and JSON Merge Patch or JSON Patch as the format. (sometimes without even knowing it!). There's hardly a clear standard, yet this is a really basic operation for any REST API.)

For a discussion of alternative crates considered, see Oxide RFD 10.

We hope Dropshot will be fairly general-purpose, but it's primarily intended to address the needs of the Oxide control plane.

Usage

The bare minimum might look like this:

use dropshot::ApiDescription;
use dropshot::ConfigDropshot;
use dropshot::ConfigLogging;
use dropshot::ConfigLoggingLevel;
use dropshot::HandlerTaskMode;
use dropshot::ServerBuilder;
use std::sync::Arc;

#[tokio::main]
async fn main() -> Result<(), String> {
    // Set up a logger.
    let log =
        ConfigLogging::StderrTerminal {
            level: ConfigLoggingLevel::Info,
        }
        .to_logger("minimal-example")
        .map_err(|e| e.to_string())?;

    // Describe the API.
    let api = ApiDescription::new();
    // Register API functions -- see detailed example or ApiDescription docs.

    // Start the server.
    let server = ServerBuilder::new(api, Arc::new(()), log)
        .start()
        .map_err(|error| format!("failed to start server: {}", error))?;

    server.await
}

This server returns a 404 for all resources because no API functions were registered. See examples/basic.rs for a simple, documented example that provides a few resources using shared state.

API Handler Functions

HTTP talks about resources. For a REST API, we often talk about endpoints or operations, which are identified by a combination of the HTTP method and the URI path.

Example endpoints for a resource called a "project" might include:

  • GET /projects (list projects)
  • POST /projects (one way to create a project)
  • GET /projects/my_project (fetch one project)
  • PUT /projects/my_project (update (or possibly create) a project)
  • DELETE /projects/my_project (delete a project)

With Dropshot, an incoming request for a given API endpoint is handled by a particular Rust function. That function is called an entrypoint, an endpoint handler, or a handler function. When you set up a Dropshot server, you configure the set of available API endpoints and which functions will handle each one by setting up an ApiDescription.

There are two ways to define a set of endpoints:

As a free function

The simplest Dropshot server defines an endpoint as a function, annotated with the endpoint macro. Here's an example of a single endpoint that lists a hardcoded project:

use dropshot::endpoint;
use dropshot::ApiDescription;
use dropshot::HttpError;
use dropshot::HttpResponseOk;
use dropshot::RequestContext;
use http::Method;
use schemars::JsonSchema;
use serde::Serialize;
use std::sync::Arc;

/// Represents a project in our API.
#[derive(Serialize, JsonSchema)]
struct Project {
    /// Name of the project.
    name: String,
}

/// Fetch a project.
#[endpoint {
    method = GET,
    path = "/projects/project1",
}]
async fn myapi_projects_get_project(
    rqctx: RequestContext<()>,
) -> Result<HttpResponseOk<Project>, HttpError>
{
   let project = Project { name: String::from("project1") };
   Ok(HttpResponseOk(project))
}

fn main() {
    let mut api = ApiDescription::new();

    // Register our endpoint and its handler function.  The "endpoint" macro
    // specifies the HTTP method and URI path that identify the endpoint,
    // allowing this metadata to live right alongside the handler function.
    api.register(myapi_projects_get_project).unwrap();

    // ... (use `api` to set up an `HttpServer` )
}

There's quite a lot going on here:

  • The endpoint macro specifies the HTTP method and URI path. When we invoke ApiDescription::register(), this information is used to register the endpoint that will be handled by our function.
  • The signature of our function indicates that on success, it returns a HttpResponseOk<Project>. This means that the function will return an HTTP 200 status code ("OK") with an object of type Project.
  • The function itself has a Rustdoc comment that will be used to document this endpoint in the OpenAPI schema.

From this information, Dropshot can generate an OpenAPI specification for this API that describes the endpoint (which OpenAPI calls an "operation"), its documentation, the possible responses that it can return, and the schema for each type of response (which can also include documentation). This is largely known statically, though generated at runtime.

As an API trait

An API trait is a Rust trait that represents a collection of API endpoints. Each endpoint is defined as a static method on the trait, and the trait as a whole is annotated with #[dropshot::api_description]. (Rust 1.75 or later is required.)

While slightly more complex than the function-based server, API traits separate the interface definition from the implementation. Keeping the definition and implementation in different crates can allow for faster iteration of the interface, and simplifies multi-service repos with clients generated from the OpenAPI output of interfaces. In addition, API traits allow for multiple implementations, such as a mock implementation for testing.

Here's an example of an API trait that's equivalent to the function-based server above:

use dropshot::ApiDescription;
use dropshot::HttpError;
use dropshot::HttpResponseOk;
use dropshot::RequestContext;
use http::Method;
use schemars::JsonSchema;
use serde::Serialize;
use std::sync::Arc;

/// Represents a project in our API.
#[derive(Serialize, JsonSchema)]
struct Project {
    /// Name of the project.
    name: String,
}

/// Defines the trait that captures all the methods.
#[dropshot::api_description]
trait ProjectApi {
    /// The context type used within endpoints.
    type Context;

    /// Fetch a project.
    #[endpoint {
        method = GET,
        path = "/projects/project1",
    }]
    async fn myapi_projects_get_project(
        rqctx: RequestContext<Self::Context>,
    ) -> Result<HttpResponseOk<Project>, HttpError>;
}

// The `dropshot::api_description` macro generates a module called
// `project_api_mod`. This module has a method called `api_description`
// that, given an implementation of the trait, returns an `ApiDescription`.
// The `ApiDescription` can then be used to set up an `HttpServer`.

// --- The following code may be in another crate ---

/// An empty type to hold the project server context.
///
/// This type is never constructed, and is purely a way to name
/// the specific server impl.
enum ServerImpl {}

impl ProjectApi for ServerImpl {
    type Context = ();

    async fn myapi_projects_get_project(
        rqctx: RequestContext<Self::Context>,
    ) -> Result<HttpResponseOk<Project>, HttpError> {
        let project = Project { name: String::from("project1") };
        Ok(HttpResponseOk(project))
    }
}

fn main() {
    // The type of `api` is provided for clarity -- it is generally inferred.
    // "api" will automatically register all endpoints defined in the trait.
    let mut api: ApiDescription<()> =
        project_api_mod::api_description::<ServerImpl>().unwrap();

    // ... (use `api` to set up an `HttpServer` )
}

See api-trait.rs and api-trait-alternate.rs for working examples.

Limitations

Currently, the #[dropshot::api_description] macro is only supported in module contexts, not within function bodies. This is a Rust limitation -- see Rust issue #79260 for more details.

Choosing between functions and traits

Prototyping: If you're prototyping with a small number of endpoints, functions provide an easier way to get started. The downside to traits is that endpoints signatures are defined at least twice, once in the trait and once in the implementation.

Small services: For a service that is relatively isolated and quick to compile, traits and functions are both good options.

APIs with multiple implementations: For services that are large enough to have a second, simpler implementation (of potentially parts of them), a trait is best.

#[endpoint { ... }] attribute parameters

The endpoint attribute accepts parameters the affect the operation of the endpoint as well as metadata that appears in the OpenAPI description of it. For more, see the documentation on endpoint.

Function parameters

In general, a handler function looks like this:

async fn f(
     rqctx: RequestContext<Context>,
     [query_params: Query<Q>,]
     [path_params: Path<P>,]
     [body_param: TypedBody<J>,]
     [body_param: UntypedBody,]
     [body_param: StreamingBody,]
     [raw_request: RawRequest,]
) -> Result<HttpResponse*, HttpError>

The RequestContext must appear first. The Context type is caller-provided context which is provided when the server is created.

The types Query, Path, TypedBody, UntypedBody, and RawRequest are called Extractors because they cause information to be pulled out of the request and made available to the handler function.

  • Query<Q> extracts parameters from a query string, deserializing them into an instance of type Q. Q must implement serde::Deserialize and schemars::JsonSchema.
  • Path<P> extracts parameters from HTTP path, deserializing them into an instance of type P. P must implement serde::Deserialize and schemars::JsonSchema.
  • TypedBody<J> extracts content from the request body by parsing the body as JSON (or form/url-encoded) and deserializing it into an instance of type J. J must implement serde::Deserialize and schemars::JsonSchema.
  • UntypedBody extracts the raw bytes of the request body.
  • StreamingBody provides the raw bytes of the request body as a Stream of Bytes chunks.
  • RawRequest provides access to the underlying hyper::Request. The hope is that this would generally not be needed. It can be useful to implement functionality not provided by Dropshot.

Query and Path impl SharedExtractor. TypedBody, UntypedBody, StreamingBody, and RawRequest impl ExclusiveExtractor. Your function may accept 0-3 extractors, but only one can be ExclusiveExtractor, and it must be the last one. Otherwise, the order of extractor arguments does not matter.

If the handler accepts any extractors and the corresponding extraction cannot be completed, the request fails with status code 400 and an error message reflecting the error (usually a validation error).

As with any serde-deserializable type, you can make fields optional by having the corresponding property of the type be an Option. Here's an example of an endpoint that takes two arguments via query parameters: "limit", a required u32, and "marker", an optional string:

use http::StatusCode;
use dropshot::HttpError;
use dropshot::TypedBody;
use dropshot::Query;
use dropshot::RequestContext;
use dropshot::Body;
use hyper::Response;
use schemars::JsonSchema;
use serde::Deserialize;
use std::sync::Arc;

#[derive(Deserialize, JsonSchema)]
struct MyQueryArgs {
    limit: u32,
    marker: Option<String>
}

struct MyContext {}

async fn myapi_projects_get(
    rqctx: RequestContext<MyContext>,
    query: Query<MyQueryArgs>)
    -> Result<Response<Body>, HttpError>
{
    let query_args = query.into_inner();
    let context: &MyContext = rqctx.context();
    let limit: u32 = query_args.limit;
    let marker: Option<String> = query_args.marker;
    Ok(Response::builder()
        .status(StatusCode::OK)
        .body(format!("limit = {}, marker = {:?}\n", limit, marker).into())?)
}

Endpoint function return types

Endpoint handler functions are async, so they always return a Future. When we say "return type" below, we use that as shorthand for the output of the future.

An endpoint function must return a type that implements HttpResponse. Typically this should be a type that implements HttpTypedResponse (either one of the Dropshot-provided ones or one of your own creation).

The more specific a type returned by the handler function, the more can be validated at build-time, and the more specific an OpenAPI schema can be generated from the source code. For example, a POST to an endpoint "/projects" might return Result<HttpResponseCreated<Project>, HttpError>. As you might expect, on success, this turns into an HTTP 201 "Created" response whose body is constructed by serializing the Project. In this example, OpenAPI tooling can identify at build time that this function produces a 201 "Created" response on success with a body whose schema matches Project (which we already said implements Serialize), and there would be no way to violate this contract at runtime.

These are the implementations of HttpTypedResponse with their associated HTTP response code on the HTTP method:

Return Type HTTP status code
HttpResponseOk 200
HttpResponseCreated 201
HttpResponseAccepted 202
HttpResponseDeleted 204
HttpResponseUpdatedNoContent 204

In situations where the response schema is not fixed, the endpoint should return Response<Body>, which also implements HttpResponse. Note that the OpenAPI spec will not include any status code or type information in this case.

What about generic handlers that run on all requests?

There's no mechanism in Dropshot for this. Instead, it's recommended that users commonize code using regular Rust functions and calling them. See the design notes in the README for more on this.

Generating OpenAPI documents

For a given ApiDescription, you can also print out an OpenAPI document describing the API. See ApiDescription::openapi.

With API traits, the #[dropshot::api_description] macro generates a helper function called stub_api_description, which returns an ApiDescription not backed by an implementation. This stub description can be used to generate an OpenAPI document for the trait without requiring an implementation of the trait. For example:

#
/// This is the API trait defined above.
#[dropshot::api_description]
trait ProjectApi {
    type Context;
    #[endpoint {
        method = GET,
        path = "/projects/project1",
    }]
    async fn myapi_projects_get_project(
        rqctx: RequestContext<Self::Context>,
    ) -> Result<HttpResponseOk<Project>, HttpError>;
}

let description = project_api_mod::stub_api_description().unwrap();
let mut openapi = description
    .openapi("Project Server", semver::Version::new(1, 0, 0));
openapi.write(&mut std::io::stdout().lock()).unwrap();

A stub description must not be used for an actual server: all request handlers will immediately panic.

Support for paginated resources

"Pagination" here refers to the interface pattern where HTTP resources (or API endpoints) that provide a list of the items in a collection return a relatively small maximum number of items per request, often called a "page" of results. Each page includes some metadata that the client can use to make another request for the next page of results. The client can repeat this until they've gotten all the results. Limiting the number of results returned per request helps bound the resource utilization and time required for any request, which in turn facilities horizontal scalability, high availability, and protection against some denial of service attacks (intentional or otherwise). For more background, see the comments in dropshot/src/pagination.rs.

Pagination support in Dropshot implements this common pattern:

  • This server exposes an API endpoint that returns the items contained within a collection.
  • The client is not allowed to list the entire collection in one request. Instead, they list the collection using a sequence of requests to the one endpoint. We call this sequence of requests a scan of the collection, and we sometimes say that the client pages through the collection.
  • The initial request in the scan may specify the scan parameters, which typically specify how the results are to be sorted (i.e., by which field(s) and whether the sort is ascending or descending), any filters to apply, etc.
  • Each request returns a page of results at a time, along with a page token that's provided with the next request as a query parameter.
  • The scan parameters cannot change between requests that are part of the same scan.
  • With all requests: there's a default limit (e.g., 100 items returned at a time). Clients can request a higher limit using a query parameter (e.g., limit=1000). This limit is capped by a hard limit on the server. If the client asks for more than the hard limit, the server can use the hard limit or reject the request.

As an example, imagine that we have an API endpoint called "/animals". Each item returned is an Animal object that might look like this:

{
    "name": "aardvark",
    "class": "mammal",
    "max_weight": "80", /* kilograms, typical */
}

There are at least 1.5 million known species of animal -- too many to return in one API call! Our API supports paginating them by "name", which we'll say is a unique field in our data set.

The first request to the API fetches "/animals" (with no querystring parameters) and returns:

{
    "next_page": "abc123...",
    "items": [
        {
            "name": "aardvark",
            "class": "mammal",
            "max_weight": "80",
        },
        ...
        {
            "name": "badger",
            "class": "mammal",
            "max_weight": "12",
        }
    ]
}

The subsequent request to the API fetches "/animals?page_token=abc123...". The page token "abc123..." is an opaque token to the client, but typically encodes the scan parameters and the value of the last item seen ("name=badger"). The client knows it has completed the scan when it receives a response with no next_page in it.

Our API endpoint can also support scanning in reverse order. In this case, when the client makes the first request, it should fetch "/animals?sort=name-descending". Now the first result might be "zebra". Again, the page token must include the scan parameters so that in subsequent requests, the API endpoint knows that we're scanning backwards, not forwards, from the value we were given. It's not allowed to change directions or sort order in the middle of a scan. (You can always start a new scan, but you can't pick up from where you were in the previous scan.)

It's also possible to support sorting by multiple fields. For example, we could support sort=class-name, which we could define to mean that we'll sort the results first by the animal's class, then by name. Thus we'd get all the amphibians in sorted order, then all the mammals, then all the reptiles. The main requirement is that the combination of fields used for pagination must be unique. We cannot paginate by the animal's class alone. (To see why: there are over 6,000 mammals. If the page size is, say, 1000, then the page_token would say "mammal", but there's not enough information there to see where we are within the list of mammals. It doesn't matter whether there are 2 mammals or 6,000 because clients can limit the page size to just one item if they want and that ought to work.)

Dropshot interfaces for pagination

The interfaces for pagination include:

  • input: your paginated API endpoint's handler function should accept an argument of type Query<PaginationParams<ScanParams, PageSelector>>, where you define ScanParams and PageSelector (see PaginationParams for more on this.)

  • output: your paginated API endpoint's handler function can return Result<HttpResponseOk<ResultsPage<T>, HttpError> where T: Serialize is the item listed by the endpoint. You can also use your own structure that contains a ResultsPage (possibly using #[serde(flatten)]), if that's the behavior you want.

See the complete, documented pagination examples in the "examples" directory for more on how to use these.

Advanced usage notes

It's possible to accept additional query parameters besides the pagination parameters by having your API endpoint handler function take two different arguments using Query, like this:

use dropshot::HttpError;
use dropshot::HttpResponseOk;
use dropshot::PaginationParams;
use dropshot::Query;
use dropshot::RequestContext;
use dropshot::ResultsPage;
use dropshot::endpoint;
use schemars::JsonSchema;
use serde::Deserialize;
use std::sync::Arc;
#[derive(Deserialize, JsonSchema)]
struct MyExtraQueryParams {
    do_extra_stuff: bool,
}

#[endpoint {
    method = GET,
    path = "/list_stuff"
}]
async fn my_list_api(
    rqctx: RequestContext<()>,
    pag_params: Query<PaginationParams<MyScanParams, MyPageSelector>>,
    extra_params: Query<MyExtraQueryParams>,
) -> Result<HttpResponseOk<ResultsPage<String>>, HttpError>
{
 # unimplemented!();
 /* ... */
}

You might expect that instead of doing this, you could define your own structure that includes a PaginationParams using #[serde(flatten)], and this ought to work, but it currently doesn't due to serde_urlencoded#33, which is really serde#1183.

Note that any parameters defined by MyScanParams are effectively encoded into the page token and need not be supplied with invocations when page_token is specified. That is not the case for required parameters defined by MyExtraQueryParams--those must be supplied on each invocation.

OpenAPI extension

In generated OpenAPI documents, Dropshot adds the x-dropshot-pagination extension to paginated operations. The value is currently a structure with this format:

{
    "required": [ .. ]
}

The string values in the required array are the names of those query parameters that are mandatory if page_token is not specified (when fetching the first page of data).

API Versioning

Dropshot servers can host multiple versions of an API. See dropshot/examples/versioning.rs for a complete, working, commented example that uses a client-provided header to determine which API version to use for each incoming request.

API versioning basically works like this:

  1. When using the endpoint macro to define an endpoint, you specify a versions field as a range of semver version strings. This identifies what versions of the API this endpoint implementation appears in. Examples:
// introduced in 1.0.0, present in all subsequent versions
versions = "1.0.0"..

// removed in 2.0.0, present in all previous versions
// (not present in 2.0.0 itself)
versions = .."2.0.0"

// introduced in 1.0.0, removed in 2.0.0
// (present only in all 1.x versions, NOT 2.0.0 or later)
versions = "1.0.0".."2.0.0"

// present in all versions (the default)
versions = ..
  1. When constructing the server, you provide VersionPolicy::Dynamic with your own impl of DynamicVersionPolicy that tells Dropshot how to determine which API version to use for each request.

  2. When a request arrives for a server using VersionPolicy::Dynamic, Dropshot uses the provided impl to determine the appropriate API version. Then it routes requests by HTTP method and path (like usual) but only considers endpoints whose version range matches the requested API version.

  3. When generating an OpenAPI document for your ApiDescription, you must provide a specific version to generate it for. It will only include endpoints present in that version and types referenced by those endpoints.

It is illegal to register multiple endpoints for the same HTTP method and path with overlapping version ranges.

All versioning-related configuration is optional. You can ignore it altogether by simply not specifying versions for each endpoint and not providing a VersionPolicy for the server (or, equivalently, providing VersionPolicy::Unversioned). In this case, the server does not try to determine a version for incoming requests. It routes requests to handlers without considering API versions.

It's maybe surprising that this mechanism only talks about versioning endpoints, but usually when we think about API versioning we think about types, especially the input and output types. This works because the endpoint implementation itself specifies the input and output types. Let's look at an example.

Suppose you have version 1.0.0 of an API with an endpoint my_endpoint with a body parameter TypedBody<MyArg>. You want to make a breaking change to the API, creating version 2.0.0 where MyArg has a new required field. You still want to support API version 1.0.0. Here's one clean way to do this:

  1. Mark the existing my_endpoint as removed after 1.0.0:
    1. Move the my_endpoint function and its input type MyArg to a new module called v1. (You'd also move its output type here if that's changing.)
    2. Change the endpoint macro invocation on my_endpoint to say versions = ..1.0.0. This says that it was removed after 1.0.0.
  2. Create a new endpoint that appears in 2.0.0.
    1. Create a new module called v2.
    2. In v2, create a new type MyArg that looks the way you want it to appear in 2.0.0. (You'd also create new versions of the output types, if those are changing, too).
    3. Also in v2, create a new my_endpoint function that accepts and returns the v2 new versions of the types. Its endpoint macro will say versions = 2.0.0.

As mentioned above, you will also need to create your server with VersionPolicy::Dynamic and specify how Dropshot should determine which version to use for each request. But that's it! Having done this:

  • If you generate an OpenAPI doc for version 1.0.0, Dropshot will include v1::my_endpoint and its types.
  • If you generate an OpenAPI doc for version 2.0.0, Dropshot will include v2::my_endpoint and its types.
  • If a request comes in for version 1.0.0, Dropshot will route it to v1::my_endpoint and so parse the body as v1::MyArg.
  • If a request comes in for version 2.0.0, Dropshot will route it to v2::my_endpoint and so parse the body as v2::MyArg.

To see a completed example of this, see dropshot/examples/versioning.rs.

DTrace probes

Dropshot optionally exposes two DTrace probes, request_start and request_finish. These provide detailed information about each request, such as their ID, the local and remote IPs, and the response information. See the dropshot::dtrace::RequestInfo and dropshot::dtrace::ResponseInfo types for a complete listing of what's available.

These probes are implemented via the usdt crate. They may require a nightly toolchain if built on macOS prior to Rust version 1.66. Otherwise a stable compiler >= v1.59 is required in order to present the necessary features. Given these constraints, USDT functionality is behind the feature flag "usdt-probes", which may become a default feature of this crate in future releases.

Important: The probes are internally registered with the DTrace kernel module, making them visible via dtrace(1M). This is done when an HttpServer object is created, but it's possible that registration fails. The result of registration is stored in the server after creation, and can be accessed with the [HttpServer::probe_registration()] method. This allows callers to decide how to handle failures, but ensures that probes are always enabled if possible.

Once in place, the probes can be seen via DTrace. For example, running:

$ cargo +nightly run --example basic --features usdt-probes

And making several requests to it with curl, we can see the DTrace probes with an invocation like:

## dtrace -Zq -n 'dropshot*:::request-* { printf("%s\n", copyinstr(arg0)); }'
{"ok":{"id":"b793c62e-60e4-45c5-9274-198a04d9abb1","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:34286","method":"GET","path":"/counter","query":null}}
{"ok":{"id":"b793c62e-60e4-45c5-9274-198a04d9abb1","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:34286","status_code":200,"message":""}}
{"ok":{"id":"9050e30a-1ce3-4d6f-be1c-69a11c618800","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:41101","method":"PUT","path":"/counter","query":null}}
{"ok":{"id":"9050e30a-1ce3-4d6f-be1c-69a11c618800","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:41101","status_code":400,"message":"do not like the number 10"}}
{"ok":{"id":"a53696af-543d-452f-81b6-5a045dd9921d","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:57376","method":"PUT","path":"/counter","query":null}}
{"ok":{"id":"a53696af-543d-452f-81b6-5a045dd9921d","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:57376","status_code":204,"message":""}}

Dependencies

~21–51MB
~1M SLoC