Learning by Doing: An HTTP API With Rust

When I started working on this post, I had another idea in mind: I wanted to compare the developer experience and performance of Spring Boot and GraalVM with Rust on a demo HTTP API application. Unfortunately, the M1 processor of my MacBook Pro had other ideas.

Tweet about the M1 processor of my MacBook Pro

Hence, I changed my initial plan. I'll write about the developer experience of developing the above application in Rust, compared to what I'm used to with Spring Boot.

The Sample Application

Like every pet project, the application is limited in scope. I designed a simple CRUD HTTP API. Data are stored in PostgreSQL.

When one designs an app on the JVM, the first and only design decision is to choose the framework: a couple of years ago, it was Spring Boot. Nowadays, the choice is mostly between Spring Boot, Quarkus, and Micronaut. In many cases, they all rely on the same underlying libraries, e.g., logging or connection pools.

Rust is much younger; hence the ecosystem has yet to mature. For every feature, one needs to choose precisely which library to use - or to implement it. Worse, one needs to understand there's such a feature. Here are the ones that I searched for:

Web Framework

The choice of the web framework is the most critical. I've to admit I had no prior clue about such libraries. I looked around and stumbled upon Which Rust web framework to choose in 2022. After reading the post, I decided to follow the conclusion and chose axum:

In particular, the last point is what sets axum apart from other frameworks. axum doesn’t have its own middleware system but instead uses tower::Service. This means axum gets timeouts, tracing, compression, authorization, and more, for free. It also enables you to share middleware with applications written using hyper or tonic.

- axum crate documentation

axum uses the Tokio asynchronous library underneath. For basic usage, it requires two crates:

TOML
 
[dependencies]
axum = "0.6"
tokio = { version = "1.23", features = ["full"] }


axum's router looks very similar to Spring's Kotlin Routes DSL:

Rust
 
let app = Router::new()
    .route("/persons", get(get_all))         //1
    .route("/persons/:id", get(get_by_id))   //1//2

async fn get_all() -> Response { ... }
async fn get_by_id(Path(id): Path<Uuid>) -> Response { ... }


  1. A route is defined by the path and a function reference.
  2. A route can have path parameters. axum can infer parameters and bind them.

Shared Objects

An issue commonly found in software projects is sharing an "object" with others. We established long ago that there were better ideas than sharing global variables.

Spring Boot (and similar JVM frameworks) solves it with runtime dependency injection. Objects are created by the framework, stored in a context, and injected into other objects when the application starts. Other frameworks do dependency injection at compile-time, e.g., Dagger 2.

Rust has neither runtime nor objects. Configurable dependency injection is not "a thing." But we can create a variable and inject it manually where needed. In Rust, it's a problem because of ownership:

Ownership is a set of rules that govern how a Rust program manages memory. All programs have to manage the way they use a computer’s memory while running. Some languages have garbage collection that regularly looks for no-longer-used memory as the program runs; in other languages, the programmer must explicitly allocate and free the memory. Rust uses a third approach: memory is managed through a system of ownership with a set of rules that the compiler checks. If any of the rules are violated, the program won’t compile. None of the features of ownership will slow down your program while it’s running.

- "What Is Ownership?"

axum provides a dedicated wrapper, the State extractor, to reuse variables across different scopes.

Rust
 
struct AppState {                                                  //1
    ...
}

impl AppState {
    fn create() -> Arc<AppState> {                                 //2
        Arc::new(AppState { ... })
    }
}

let app_state = AppState::create();
let app = Router::new()
    .route("/persons", get(get_all))
    .with_state(Arc::clone(&app_state));                           //3

async fn get_all(State(state): State<Arc<AppState>>) -> Response { //4
    ...                                                            //5
}


  1. Create the struct to be shared.
  2. Create a new struct wrapped in an Atomically Reference Counted.
  3. Share the reference with all routing functions, e.g., get_all
  4. Pass the state
  5. Use it!

Automated JSON Serialization

Modern JVM web frameworks automatically serialize objects in JSON before sending. The good thing is that axum does the same. It relies on Serde. First, we add the serde and serde_json crate dependencies:

TOML
 
[dependencies]
serde = { version = "1.0", features = ["derive"] }
serde_json = "1.0"


Then, we annotate our struct with the derive(Serialize) macro:

Rust
 
#[derive(Serialize)]
struct Person {
    first_name: String,
    last_name: String,
}


Finally, we return the struct wrapped in a Json and the HTTP status code in an axum Response.

Rust
 
async fn get_test() -> impl IntoResponse {        //1
    let person = Person {                         //2
        first_name: "John".to_string(),
        last_name: "Doe".to_string()
    };
    (StatusCode::OK, Json(person))                //3
}


  1. The tuple (StatusCode, Json) is automatically converted into a Response
  2. Create the Person
  3. Return the tuple.

At runtime, axum automatically serializes the struct in JSON:

JSON
 
{"first_name":"Jane","last_name":"Doe"}


Database Access

For a long time, I used the MySQL database for my demos, but I started to read a lot of good stuff about PostgreSQL and decided to switch. I needed an asynchronous library compatible with Tokio: it's exactly what the tokio_postgres crate does.

The problem with the crate is that it creates direct connections to the database. I searched for a connection pool crate and stumbled upon deadpool (sic):

Deadpool is a dead simple async pool for connections and objects of any type.

- Deadpool

Deadpool provides two distinct implementations:

More specialized implementations of the latter cater to different databases or "drivers", e.g., Redis and... tokio-postgres. One can configure Deadpool directly or defer to the config crate it supports. The latter crate allows several alternatives for configuration:

Config organizes hierarchical or layered configurations for Rust applications.

Config lets you set a set of default parameters and then extend them via merging in configuration from a variety of sources:

Additionally, Config supports:

- Crate config

To create the base configuration, one needs to create a dedicated structure and use the crate:

Rust
 
#[derive(Deserialize)]                                       //1
struct ConfigBuilder {
    postgres: deadpool_postgres::Config,                     //2
}

impl ConfigBuilder {
    async fn from_env() -> Result<Self, ConfigError> {       //3
        Config::builder()
            .add_source(
                Environment::with_prefix("POSTGRES")         //4
                    .separator("_")                          //4
                    .keep_prefix(true)                       //5
                    .try_parsing(true),
            )
            .build()?
            .try_deserialize()
    }
}

let cfg_builder = ConfigBuilder::from_env().await.unwrap();  //6


  1. The Deserialize macro is mandatory.
  2. The field must match the environment prefix (see below).
  3. The function is async and returns a Result
  4. Read from environment variables whose name starts with POSTGRES_
  5. Keep the prefix in the configuration map. 
  6. Enjoy!

Note that environment variables should conform to what Deadpool's Config expects. Here's my configuration in Docker Compose:

Env variable Value
POSTGRES_HOST "postgres"
POSTGRES_PORT 5432
POSTGRES_USER "postgres"
POSTGRES_PASSWORD "root"
POSTGRES_DBNAME "app"

Once we have initialized the configuration, we can create the pool:

Rust
 
struct AppState {
    pool: Pool,                                                     //1
}

impl AppState {
    async fn create() -> Arc<AppState> {                            //2
        let cfg_builder = ConfigBuilder::from_env().await.unwrap(); //3
        let pool = cfg_builder                                      //4
            .postgres
            .create_pool(
                Some(deadpool_postgres::Runtime::Tokio1),
                tokio_postgres::NoTls,
            )
            .unwrap();
        Arc::new(AppState { pool })                                 //2
    }
}


  1. Wrap the pool in a custom struct
  2. Wrap the struct in an Arc to pass it within an axumState (see above). 
  3. Get the configuration.
  4. Create the pool.

Then, we can pass the pool to the routing functions:

Rust
 
let app_state = AppState::create().await;                           //1
let app = Router::new()
    .route("/persons", get(get_all))
    .with_state(Arc::clone(&app_state));                            //2

async fn get_all(State(state): State<Arc<AppState>>) -> Response {
    let client = state.pool.get().await.unwrap();                   //3
    let rows = client
        .query("SELECT id, first_name, last_name FROM person", &[]) //4
        .await                                                      //5
        .unwrap();
    //                                                              //6
}


  1. Create the state.
  2. Pass the state to the routing functions.
  3. Get the pool out of the state, and get the client out of the pool.
  4. Create the query.
  5. Execute it.
  6. Read the row to populate the Response

The last step is to implement the transformation from a Row to a Person. We can do it with the From trait.

Rust
 
impl From<&Row> for Person {
    fn from(row: &Row) -> Self {
        let first_name: String = row.get("first_name");
        let last_name: String = row.get("last_name");
        Person {
            first_name,
            last_name,
        }
    }
}

let person = row.into();


Docker Build

The last step is the building of the application. I want everybody to be able to build, so I used Docker. Here's the Dockerfile:

Dockerfile
 
FROM --platform=x86_64 rust:1-slim AS build                                  //1

RUN rustup target add x86_64-unknown-linux-musl                              //2
RUN apt update && apt install -y musl-tools musl-dev                         //3

WORKDIR /home

COPY Cargo.toml .
COPY Cargo.lock .
COPY src src

RUN --mount=type=cache,target=/home/.cargo \                                 //4
 && cargo build --target x86_64-unknown-linux-musl --release                 //5

FROM scratch                                                                 //6

COPY --from=build /home/target/x86_64-unknown-linux-musl/release/rust /app   //7

CMD ["/app"]


  1. Start from a standard Rust image.
  2. Add musl target so we can compile to Alpine Linux.
  3. Install the required Alpine dependencies.
  4. Cache the dependencies.
  5. Build for Alpine Linux.
  6. Start from scratch.
  7. Add the previously built binary.

The final image is 7.56MB. My experience has shown that an equivalent GraalVM native compiled image would be more than 100MB.

Conclusion

Though it was not my initial plan, I learned about quite a few libraries with this demo app and how they work.

More importantly, I've experienced what it is like to develop an app without a framework like Spring Boot. You need to know the following:

  1. Available crates for each capability
  2. Crate compatibility
  3. Version compatibility

Last but not least, the documentation of most above crates ranges from average to good. I found axum's to be good; on the other hand, I didn't manage to use Deadpool correctly from the start and had to go through several iterations. The documentation quality of Rust crates is different from crate to crate. All in all, they have room for the potential to reach the level of modern JVM frameworks.

Also, the demo app was quite simple. I assume that more advanced features could be more painful.

The complete source code for this post can be found on GitHub.

To go further:

 

 

 

 

Top