Cogs and Levers A blog full of technical stuff

Learning Rust Part 5 - Data Structures

Introduction

Rust offers a versatile range of data structures that make working with various types of data efficient and safe. In this post, we’ll cover fundamental data structures like vectors, hash maps, and arrays, along with Rust’s powerful enums and structs for creating custom data types.

Text

Strings and String Slices

Rust provides two main string types:

  • String: A growable, heap-allocated string.
  • &str (string slice): An immutable view into a string, commonly used for read-only access.

String example:

let mut s = String::from("Hello");
s.push_str(", world!");
println!("{}", s);

String slice example:

let greeting = "Hello, world!";
let slice = &greeting[0..5]; // "Hello"

Collections

Vectors

Vectors (Vec<T>) are dynamic arrays that can grow or shrink in size, making them ideal for storing sequences of values when the size isn’t known at compile-time.

fn main() {
    let mut numbers = vec![1, 2, 3];
    numbers.push(4); // Add an element
    println!("{:?}", numbers);
}

LinkedList

A LinkedList<T> is a doubly linked list that allows fast insertion and deletion of elements at both ends of the list. It is less commonly used than Vec but can be useful when you need to insert and remove elements frequently from both the front and back of the collection.

 
use std::collections::LinkedList;

fn main() { 
    let mut list = LinkedList::new(); 

    list.push_back(1); 
    list.push_back(2); 
    list.push_front(0);

    for value in &list {
        println!("{}", value); // Prints 0, 1, 2
    }
} 

HashMap

A HashMap<K, V> stores key-value pairs, enabling efficient value retrieval based on keys.

use std::collections::HashMap;

fn main() {
    let mut scores = HashMap::new();
    scores.insert("Alice", 10);
    scores.insert("Bob", 20);

    println!("{:?}", scores.get("Alice")); // Some(&10)
}

BTreeMap

BTreeMap<K, V> is similar to HashMap but keeps keys sorted, making it useful for scenarios where sorted keys are necessary.

use std::collections::BTreeMap;

fn main() {
    let mut scores = BTreeMap::new();
    scores.insert("Alice", 10);
    scores.insert("Bob", 20);

    for (key, value) in &scores {
        println!("{}: {}", key, value);
    }
}

BinaryHeap

A BinaryHeap<T> is a priority queue implemented with a binary heap, where elements are always ordered so the largest (or smallest) element can be accessed quickly. By default, BinaryHeap maintains a max-heap, but it can be customized for min-heap operations.

 
use std::collections::BinaryHeap;

fn main() { 
    let mut heap = BinaryHeap::new(); 

    heap.push(1); 
    heap.push(5); 
    heap.push(3);

    while let Some(top) = heap.pop() {
        println!("{}", top); // Prints values in descending order: 5, 3, 1
    }
}

HashSet

A HashSet<T> is a collection of unique values, implemented as a hash table. It provides fast membership checking and is useful when you need to store non-duplicate items without any specific order.

 
use std::collections::HashSet;

fn main() { 
    let mut set = HashSet::new(); 

    set.insert("apple"); 
    set.insert("banana"); 
    set.insert("apple"); // Duplicate, ignored by the set

    println!("{:?}", set.contains("banana")); // true
    println!("{:?}", set); // {"apple", "banana"}
}

BTreeSet

A BTreeSet<T> is a sorted, balanced binary tree-based set. Like HashSet, it only stores unique values, but unlike HashSet, it maintains its items in sorted order. This makes it suitable for range queries and ordered data.

 
use std::collections::BTreeSet;

fn main() { 
    let mut set = BTreeSet::new(); 

    set.insert(10); 
    set.insert(20); 
    set.insert(15);

    for value in &set {
        println!("{}", value); // Prints 10, 15, 20 in sorted order
    }
}

VecDeque

A VecDeque<T> (double-ended queue) is a resizable, efficient data structure that supports adding and removing elements from both the front and back. It’s ideal for queue-like operations where both ends need to be accessible.

 
use std::collections::VecDeque;

fn main() { 
    let mut deque = VecDeque::new(); 

    deque.push_back(1); 
    deque.push_front(0);

    println!("{:?}", deque.pop_back()); // Some(1)
    println!("{:?}", deque.pop_front()); // Some(0)

} 

Option and Result Types

Rust’s Option and Result types are enums that enable safe handling of optional values and errors. We went over error handling in part 3 of this series.

Option

The Option<T> type represents an optional value: Some(T) for a value, or None if absent.

fn find_element(index: usize) -> Option<i32> {
    let numbers = vec![1, 2, 3];
    numbers.get(index).copied()
}

fn main() {
    match find_element(1) {
        Some(number) => println!("Found: {}", number),
        None => println!("Not found"),
    }
}

Result

The Result<T, E> type is used for functions that may succeed (Ok(T)) or fail (Err(E)), promoting explicit error handling.

fn divide(a: f64, b: f64) -> Result<f64, &'static str> {
    if b == 0.0 {
        Err("Cannot divide by zero")
    } else {
        Ok(a / b)
    }
}

Custom Data Types and Enums

Rust’s enums and structs allow you to create custom data types, essential for building complex and expressive programs.

Enums

Rust’s enums can hold different types of data within each variant, enabling versatile data representations.

enum Message {
    Text(String),
    Image { url: String, width: u32, height: u32 },
    Quit,
}

fn main() {
    let msg = Message::Text(String::from("Hello"));
    match msg {
        Message::Text(text) => println!("Text: {}", text),
        Message::Image { url, width, height } => println!("Image at {}, size: {}x{}", url, width, height),
        Message::Quit => println!("Quit message"),
    }
}

Structs and Tuple Structs

Structs allow for creating complex types with named fields.

struct Person {
    name: String,
    age: u8,
}

fn main() {
    let person = Person {
        name: String::from("Alice"),
        age: 30,
    };
    println!("Name: {}, Age: {}", person.name, person.age);
}

Tuple structs are useful for grouping values without naming fields, often for simpler data types.

struct Color(u8, u8, u8);

fn main() {
    let red = Color(255, 0, 0);
    println!("Red: {}, {}, {}", red.0, red.1, red.2);
}

Arrays, Slices, and Compile-Time Length Arrays

Arrays

Arrays in Rust are fixed-size collections of elements with known length at compile-time. They’re stack-allocated, offering efficiency and safety.

fn main() {
    let numbers: [i32; 3] = [1, 2, 3];
    println!("First element: {}", numbers[0]);
}

Slices

Slices provide a way to view sections of an array or vector, avoiding the need to copy data.

fn main() {
    let numbers = [1, 2, 3, 4];
    let slice = &numbers[1..3];
    println!("{:?}", slice); // [2, 3]
}

Slices work with both arrays and vectors and are typically used as function parameters to avoid copying large data structures.

Reference Counting

Rc<Vec<T>> and Arc<Vec<T>>

Rc (Reference Counting) and Arc (Atomic Reference Counting) are common wrappers around collections like Vec to allow multiple ownership. Rc is single-threaded, while Arc is thread-safe, and both are used frequently for sharing collections between parts of a program.

 
use std::rc::Rc; 
use std::sync::Arc;

fn main() { 
    let vec = vec![1, 2, 3]; 
    let shared_rc = Rc::new(vec.clone()); 
    let shared_arc = Arc::new(vec);

    println!("Rc count: {}", Rc::strong_count(&shared_rc)); // Rc count: 1
    println!("Arc count: {}", Arc::strong_count(&shared_arc)); // Arc count: 1

} 

Summary

Rust’s data structures—from collections like Vec and HashMap to custom types with struct and enum—enable flexible, efficient, and safe data handling. With tools like Option and Result, Rust enforces a safety-first approach without compromising on performance, making it an ideal language for robust application development.

Learning Rust Part 4 - Concurrency

Introduction

Rust’s concurrency model provides a unique approach to safe parallel programming by eliminating data races and encouraging structured, reliable concurrent code. Through its ownership model, concurrency primitives, and async/await syntax, Rust enables developers to write efficient, parallel programs. In this post, we’ll explore Rust’s key tools and patterns for safe concurrency.

Threads and Thread Safety

Rust’s std::thread module allows developers to create threads, enabling programs to perform multiple tasks concurrently.

Creating Threads

Rust threads are created with std::thread::spawn, and they can run independently of the main thread. The join method is used to wait for threads to complete.

use std::thread;

fn main() {
    let handle = thread::spawn(|| {
        for i in 1..5 {
            println!("Thread: {}", i);
        }
    });

    for i in 1..5 {
        println!("Main: {}", i);
    }

    handle.join().unwrap(); // Wait for the thread to finish
}

Thread Safety

Rust’s ownership model ensures that data shared across threads is managed safely. Rust achieves this through two primary mechanisms:

  • Ownership Transfer: Data can be transferred to threads, where the original owner relinquishes control.
  • Immutable Sharing: If data is borrowed immutably, it can be accessed concurrently across threads without modification.

Concurrency Primitives (Mutex, RwLock)

Rust offers concurrency primitives, such as Mutex and RwLock, to allow safe mutable data sharing across threads.

Mutex (Mutual Exclusion)

A Mutex ensures that only one thread can access the data at a time. When using lock() on a Mutex, it returns a guard that releases the lock automatically when dropped.

use std::sync::{Mutex, Arc};
use std::thread;

fn main() {
    let data = Arc::new(Mutex::new(0));
    let mut handles = vec![];

    for _ in 0..10 {
        let data = Arc::clone(&data);
        let handle = thread::spawn(move || {
            let mut num = data.lock().unwrap();
            *num += 1;
        });
        handles.push(handle);
    }

    for handle in handles {
        handle.join().unwrap();
    }

    println!("Result: {}", *data.lock().unwrap());
}

RwLock (Read-Write Lock)

An RwLock allows multiple readers or a single writer, making it ideal for scenarios where data is read often but updated infrequently.

use std::sync::{RwLock, Arc};

fn main() {
    let data = Arc::new(RwLock::new(0));

    {
        let read_data = data.read().unwrap();
        println!("Read: {}", *read_data);
    }

    {
        let mut write_data = data.write().unwrap();
        *write_data += 1;
    }
}

Atomic Types

Atomic types like AtomicBool, AtomicIsize, and AtomicUsize enable lock-free, atomic operations on shared data, which is useful for simple counters or flags.

use std::sync::atomic::{AtomicUsize, Ordering};
use std::thread;

fn main() {
    let counter = AtomicUsize::new(0);

    let handles: Vec<_> = (0..10).map(|_| {
        thread::spawn(|| {
            counter.fetch_add(1, Ordering::SeqCst);
        })
    }).collect();

    for handle in handles {
        handle.join().unwrap();
    }

    println!("Counter: {}", counter.load(Ordering::SeqCst));
}

Channel Communication

Rust’s channels, provided by the std::sync::mpsc module, allow message passing between threads. Channels provide safe communication without shared memory, following a multiple-producer, single-consumer pattern.

Creating and Using Channels

use std::sync::mpsc;
use std::thread;

fn main() {
    let (tx, rx) = mpsc::channel();

    thread::spawn(move || {
        let message = String::from("Hello from thread");
        tx.send(message).unwrap();
    });

    let received = rx.recv().unwrap();
    println!("Received: {}", received);
}

Multi-Threaded Producers

To enable multiple threads to send messages to the same receiver, you can clone the transmitter.

let tx = mpsc::Sender::clone(&tx);

Async/Await and Asynchronous Programming

Rust’s async/await syntax supports asynchronous programming, allowing tasks to pause (await) without blocking the entire thread. Async functions in Rust return Future types, which represent values available at a later time.

Defining and Using Async Functions

An async function returns a Future and only runs when awaited.

async fn fetch_data() -> u32 {
    42
}

#[tokio::main]
async fn main() {
    let data = fetch_data().await;
    println!("Data: {}", data);
}

.await will force the application to wait for fetch_data() to complete before moving on.

Combining Async Functions

Multiple async calls can be combined with tokio::join!, allowing concurrency without additional threads.

async fn first() -> u32 { 10 }
async fn second() -> u32 { 20 }

async fn run() {
    let (a, b) = tokio::join!(first(), second());
    println!("Sum: {}", a + b);
}

Task-Based Concurrency with Tokio and async-std

Rust offers runtime libraries like Tokio and async-std for task-based concurrency, providing asynchronous runtimes suited for managing complex async workflows.

Using Tokio

Tokio is a popular async runtime, offering tools for task management, timers, and network I/O.

use tokio::time::{sleep, Duration};

#[tokio::main]
async fn main() {
    let handle = tokio::spawn(async {
        sleep(Duration::from_secs(1)).await;
        println!("Task completed!");
    });

    handle.await.unwrap();
}

async-std Example

async-std offers similar functionality with a simpler API for certain tasks.

use async_std::task;
use std::time::Duration;

fn main() {
    task::block_on(async {
        task::sleep(Duration::from_secs(1)).await;
        println!("Task completed!");
    });
}

Summary

Rust’s concurrency model provides robust tools for safe multithreading and asynchronous programming. By combining threads, async/await syntax, and concurrency primitives like Mutex and RwLock, Rust enables safe data sharing and task-based concurrency, making it a powerful choice for high-performance concurrent applications.

Learning Rust Part 16 - Interoperability

Introduction

Rust’s FFI (Foreign Function Interface) capabilities and rich library support enable it to integrate seamlessly with other languages like C, C++, and Python. Rust can also produce shared libraries and handle various data interchange formats such as JSON, Protobuf, and MsgPack, making it a great choice for cross-language applications and APIs. This post covers essential tools and techniques for interfacing Rust with other languages.

FFI with C and C++

Rust’s FFI makes it possible to interact directly with C libraries, letting Rust leverage existing C code or integrate with languages like C++. The extern keyword and the libc crate facilitate this interoperability.

Calling C Functions from Rust

To call a C function, define an extern block and use #[link] to specify the library. Here’s an example with the C sqrt function from the math library:

extern "C" {
    fn sqrt(x: f64) -> f64;
}

fn main() {
    let x = 25.0;
    unsafe {
        println!("sqrt({}) = {}", x, sqrt(x));
    }
}

Exposing Rust Functions to C

To expose Rust functions for use in C, use #[no_mangle] and declare the function as extern "C". This prevents Rust from altering the function name.

#[no_mangle]
pub extern "C" fn rust_add(a: i32, b: i32) -> i32 {
    a + b
}

Interfacing with C++ using the cxx crate

The cxx crate provides an interface for calling C++ code from Rust and vice versa, handling C++ types like std::string and std::vector.

Add cxx to Cargo.toml and define a C++ bridge file (bridge.rs):

#[cxx::bridge]
mod ffi {
    extern "C++" {
        include!("example.h");
        fn cpp_function(x: i32) -> i32;
    }
}

fn main() {
    let result = ffi::cpp_function(42);
    println!("Result from C++: {}", result);
}

Rust and Python Interfacing with pyo3

The pyo3 crate allows Rust to execute Python code, call Python functions, and even create Python modules directly from Rust.

Calling Python Code from Rust

Use pyo3 to execute Python code within Rust. First, add pyo3 to Cargo.toml:

[dependencies]
pyo3 = { version = "0.15", features = ["extension-module"] }

Then, write a Rust function that interacts with Python:

use pyo3::prelude::*;

fn main() -> PyResult<()> {
    Python::with_gil(|py| {
        let sys = py.import("sys")?;
        let version: String = sys.get("version")?.extract()?;
        println!("Python version: {}", version);
        Ok(())
    })
}

Building a Python Module in Rust

Rust can also create native Python modules. Annotate functions with #[pyfunction] and use #[pymodule] to define the module.

use pyo3::prelude::*;

#[pyfunction]
fn sum_as_string(a: i64, b: i64) -> PyResult<String> {
    Ok((a + b).to_string())
}

#[pymodule]
fn my_rust_module(py: Python, m: &PyModule) -> PyResult<()> {
    m.add_function(wrap_pyfunction!(sum_as_string, m)?)?;
    Ok(())
}

Build this as a shared library, and it can be imported into Python just like a native module.

Building Shared Libraries

Rust can produce shared libraries (e.g., .dll on Windows, .so on Linux, and .dylib on macOS), making it easy to share Rust code across multiple languages.

Compiling Rust to a Shared Library

To build a Rust project as a shared library, set the crate-type in Cargo.toml:

[lib]
crate-type = ["cdylib"]

Then build the library with:

cargo build --release

This generates a .dll, .so, or .dylib file, depending on your operating system, which other languages can link to and use.

Using the Shared Library

From another language, import the shared library and call its functions. For instance, in Python, you can use ctypes to load and call functions from the Rust shared library:

import ctypes

lib = ctypes.CDLL('./target/release/libmy_rust_lib.so')
result = lib.rust_add(10, 20)
print(f"Result from Rust: {result}")

Using Rust with Other Languages

Rust can interface with languages like JavaScript, Ruby, and Go by using FFI or compiling Rust to WebAssembly or shared libraries.

WebAssembly (Wasm) for JavaScript Interoperability

WebAssembly allows Rust code to run in the browser or JavaScript environments. Using wasm-bindgen, Rust functions can be exposed to JavaScript.

Add wasm-bindgen to Cargo.toml:

use wasm_bindgen::prelude::*;

#[wasm_bindgen]
pub fn greet(name: &str) -> String {
    format!("Hello, {}!", name)
}

Build the Rust code as WebAssembly and import it in JavaScript, making Rust interoperable with frontend applications.

Data Interchange Formats (JSON, Protobuf, MsgPack)

Rust supports serialization formats that allow data interchange with other systems and languages.

JSON with serde_json

The serde_json crate is the standard for JSON serialization and deserialization in Rust.

use serde::{Serialize, Deserialize};
use serde_json;

#[derive(Serialize, Deserialize)]
struct User {
    id: u32,
    name: String,
}

fn main() -> serde_json::Result<()> {
    let user = User { id: 1, name: "Alice".to_string() };
    let json = serde_json::to_string(&user)?;
    println!("Serialized JSON: {}", json);

    let deserialized: User = serde_json::from_str(&json)?;
    println!("Deserialized: {:?}", deserialized);
    Ok(())
}

Protobuf with prost

Google’s Protocol Buffers (Protobuf) is a fast, language-agnostic format used for efficient data serialization. Rust’s prost crate generates Rust types from .proto files.

Define a .proto file for your data structures and use prost to generate Rust types.

use prost::Message;

#[derive(Message)]
struct User {
    #[prost(uint32, tag = "1")]
    pub id: u32,
    #[prost(string, tag = "2")]
    pub name: String,
}

MsgPack with rmp-serde

MsgPack is a compact, binary format for data serialization, providing efficiency for high-performance applications. rmp-serde allows Rust to serialize and deserialize MsgPack data using serde.

use serde::{Serialize, Deserialize};
use rmp_serde::{to_vec, from_slice};

#[derive(Serialize, Deserialize, Debug)]
struct User {
    id: u32,
    name: String,
}

fn main() -> Result<(), Box<dyn std::error::Error>> {
    let user = User { id: 1, name: "Alice".to_string() };
    let msgpack = to_vec(&user)?; // Serialize to MsgPack

    let deserialized: User = from_slice(&msgpack)?; // Deserialize
    println!("Deserialized: {:?}", deserialized);
    Ok(())
}

Summary

Rust’s interoperability capabilities make it ideal for building cross-language applications. Whether through FFI, shared libraries, or data interchange formats like JSON and Protobuf, Rust can integrate seamlessly with various ecosystems, enabling it to act as a high-performance backend or computational layer in multi-language projects.

Learning Rust Part 15 - Systems Programming

Introduction

Rust’s combination of low-level control, memory safety, and performance makes it an excellent choice for systems programming. Rust supports direct memory management, OS interfacing, and embedded programming while minimizing undefined behavior. In this post, we’ll explore essential systems programming topics, including memory management, device driver development, and embedded systems programming.

Low-level Memory Management

Rust’s memory management model enforces safe practices without a garbage collector. Tools like Box, Rc, Arc, and unsafe allow for direct memory management.

Using Box for Heap Allocation

Box<T> is used for heap allocation, ideal for large data structures that may not fit on the stack. By default, Rust allocates on the stack, but Box moves data to the heap.

fn main() {
    let boxed_value = Box::new(10);
    println!("Boxed value: {}", boxed_value);
}

Unsafe Rust for Manual Memory Management

Rust ensures safety by default, but unsafe blocks enable direct memory access, pointer manipulation, and interfacing with other languages, useful for hardware interactions or optimizing critical code paths.

fn unsafe_memory() {
    let x = 10;
    let r = &x as *const i32;

    unsafe {
        println!("Unsafe pointer dereference: {}", *r);
    }
}

Interfacing with Operating System APIs

Rust’s std::os and libc crates provide access to OS-specific APIs, enabling low-level system calls, process management, and file descriptor handling.

Working with Files and File Descriptors

While std::fs handles files at a high level, std::os::unix and std::os::windows provide OS-specific functionality for working with raw file descriptors.

use std::os::unix::io::{RawFd, AsRawFd};
use std::fs::File;

fn main() -> std::io::Result<()> {
    let file = File::open("example.txt")?;
    let raw_fd: RawFd = file.as_raw_fd();
    println!("Raw file descriptor: {}", raw_fd);
    Ok(())
}

Calling OS Functions with libc

The libc crate allows calling C library functions directly, giving access to various POSIX functions for low-level system programming.

extern crate libc;
use libc::{getpid, c_int};

fn main() {
    let pid: c_int = unsafe { getpid() };
    println!("Process ID: {}", pid);
}

Writing Device Drivers

Rust is increasingly popular for device drivers because of its safety guarantees. While driver development requires unsafe code to interact directly with hardware, Rust’s borrow checker reduces common errors.

Example: Writing a Basic Character Device Driver

Creating an actual device driver requires interacting with kernel space. Below is a basic structure that mimics a character device driver.

#![no_std]
#![no_main]

extern crate embedded_hal as hal;
use hal::blocking::serial::Write;
use core::fmt::Write as FmtWrite;

struct Serial;

impl Write<u8> for Serial {
    type Error = ();

    fn bwrite_all(&mut self, buffer: &[u8]) -> Result<(), Self::Error> {
        for &byte in buffer {
            unsafe { core::ptr::write_volatile(0x4000_0000 as *mut u8, byte) };
        }
        Ok(())
    }
}

This sample initializes a Serial struct to write directly to a memory-mapped I/O address.

Embedded Systems with no_std

Rust’s no_std environment enables development without the standard library, essential for embedded systems where resources are limited. In no_std projects, libraries like embedded-hal provide low-level functionalities for microcontrollers.

Creating a no_std Embedded Project

To work in an embedded environment, first disable the standard library by specifying #![no_std]. Libraries like cortex-m and embedded-hal provide core functionalities for microcontrollers.

#![no_std]
#![no_main]

extern crate cortex_m_rt as rt;
use rt::entry;

#[entry]
fn main() -> ! {
    // Your embedded code here
    loop {}
}

The #[entry] macro designates the entry point, while #![no_std] removes the dependency on the standard library.

Building Kernels and Operating Systems

Rust is becoming popular for experimental operating systems and kernel development due to its safety and performance. Kernel development in Rust uses no_std, allowing low-level hardware control.

Example Structure for a Basic OS Kernel

To create a basic OS kernel, use #![no_std] and #![no_main] with a custom entry point, typically _start. Since the standard library is unavailable, you handle everything directly with low-level code.

#![no_std]
#![no_main]

use core::panic::PanicInfo;

#[no_mangle]
pub extern "C" fn _start() -> ! {
    loop {}
}

#[panic_handler]
fn panic(_info: &PanicInfo) -> ! {
    loop {}
}

This code provides a minimal structure for a Rust-based OS kernel, with _start as the entry point and a custom panic_handler.

Performance Optimizations and Profiling

Rust offers various tools for profiling and optimizing performance, including compiler flags, profiling tools, and benchmarking libraries like criterion.

Compiler Flags for Optimization

Using cargo build --release enables optimizations, significantly improving performance by enabling Rust’s optimization passes.

cargo build --release

Profiling with perf

For detailed profiling, Rust projects can use perf on Linux to gain insights into CPU usage and performance bottlenecks.

Compile with Release Mode

cargo build --release

Run with perf

perf record ./target/release/your_binary
perf report

Criterion for Benchmarking

criterion is a Rust library for benchmarking, providing reliable and statistically sound measurements for performance testing.

use criterion::{black_box, criterion_group, criterion_main, Criterion};

fn fibonacci(n: u64) -> u64 {
    match n {
        0 => 1,
        1 => 1,
        n => fibonacci(n - 1) + fibonacci(n - 2),
    }
}

fn criterion_benchmark(c: &mut Criterion) {
    c.bench_function("fib 20", |b| b.iter(|| fibonacci(black_box(20))));
}

criterion_group!(benches, criterion_benchmark);
criterion_main!(benches);

Run with cargo bench to get detailed performance data.

Summary

Rust’s systems programming capabilities make it an exceptional tool for low-level development. With control over memory, access to OS APIs, support for embedded systems, and tools for profiling and optimization, Rust combines safety and performance, enabling a wide range of system-level applications.

Learning Rust Part 14 - Security and Cryptography

Introduction

Rust’s strong memory safety guarantees and growing ecosystem of security libraries make it an excellent choice for building secure applications. From encryption and password hashing to secure communication and cross-compilation for secure systems, Rust provides a solid foundation for high-security applications. In this post, we’ll explore key tools and techniques for secure application development in Rust.

Encryption Libraries (e.g., rust-crypto, ring)

Rust offers a range of encryption libraries, including rust-crypto and ring, which provide cryptographic algorithms like AES, RSA, and SHA-2. These libraries enable secure encryption, decryption, hashing, and digital signatures.

Using ring for Encryption and Hashing

The ring library is popular for cryptographic operations in Rust, offering efficiency and ease of use.

Example: Hashing with SHA-256

use ring::digest::{Context, SHA256, Digest};

fn sha256_hash(data: &[u8]) -> Digest {
    let mut context = Context::new(&SHA256);
    context.update(data);
    context.finish()
}

fn main() {
    let data = b"Hello, Rust!";
    let hash = sha256_hash(data);
    println!("SHA-256 hash: {:?}", hash);
}

Example: AES Encryption and Decryption with ring

The ring library provides AES-GCM for authenticated encryption, ensuring both confidentiality and data integrity.

use ring::aead::{AES_256_GCM, SealingKey, UnboundKey, LessSafeKey, Nonce, Aad, OpeningKey};
use ring::rand::{SystemRandom, SecureRandom};

fn aes_encrypt(data: &[u8], key: &[u8]) -> Vec<u8> {
    let unbound_key = UnboundKey::new(&AES_256_GCM, key).unwrap();
    let sealing_key = LessSafeKey::new(unbound_key);
    let nonce = Nonce::assume_unique_for_key([0u8; 12]);
    let mut in_out = data.to_vec();
    sealing_key.seal_in_place_append_tag(nonce, Aad::empty(), &mut in_out).unwrap();
    in_out
}

Password Hashing and Secure Storage

Password hashing is crucial for securely storing user passwords. Libraries like argon2 provide key derivation functions (e.g., Argon2, scrypt, bcrypt) that are secure against brute-force attacks.

Argon2 with argon2 Library

The argon2 crate enables secure password hashing, an essential feature for storing user credentials.

use argon2::{self, Config};

fn hash_password(password: &[u8], salt: &[u8]) -> Vec<u8> {
    let config = Config::default();
    argon2::hash_raw(password, salt, &config).unwrap()
}

fn main() {
    let password = b"super_secret_password";
    let salt = b"random_salt";
    let hash = hash_password(password, salt);
    println!("Password hash: {:?}", hash);
}

Secure Storage

Storing sensitive information, like API keys and secrets, securely is essential. You can use encrypted databases or dedicated secure storage libraries like secrecy to ensure data stays confidential in memory.

TLS and SSL with rustls

For secure communication, Rust provides rustls, a TLS library built on ring. Unlike C-based libraries like OpenSSL, rustls is memory-safe and avoids common vulnerabilities.

Setting up a TLS Server with rustls

Using rustls, you can build a TLS-enabled server that ensures secure data transmission.

use rustls::{ServerConfig, NoClientAuth};
use std::sync::Arc;
use tokio_rustls::TlsAcceptor;
use tokio::net::TcpListener;

#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
    let certs = load_certs("cert.pem")?;
    let key = load_private_key("key.pem")?;
    let config = ServerConfig::builder()
        .with_safe_defaults()
        .with_no_client_auth()
        .with_single_cert(certs, key)?;

    let listener = TcpListener::bind("127.0.0.1:8443").await?;
    let acceptor = TlsAcceptor::from(Arc::new(config));

    loop {
        let (socket, _) = listener.accept().await?;
        let acceptor = acceptor.clone();
        tokio::spawn(async move {
            let _tls_stream = acceptor.accept(socket).await.unwrap();
            println!("TLS connection established");
        });
    }
}

In this example, rustls is configured with certificates for server authentication, and incoming connections are wrapped in TLS for secure communication.

Cross-compilation for Secure Systems

Cross-compilation allows you to build Rust applications for secure or embedded environments, such as ARM-based systems or Linux-based IoT devices. Tools like rustup and custom target configurations facilitate cross-compiling Rust code.

Example: Cross-compiling for ARM

To cross-compile for an ARM-based system (e.g., Raspberry Pi), use rustup to install the appropriate target.

rustup target add armv7-unknown-linux-gnueabihf
cargo build --target armv7-unknown-linux-gnueabihf

For more secure systems, you can use musl as a static linking target, ensuring binary compatibility and reducing dependencies.

rustup target add x86_64-unknown-linux-musl
cargo build --target x86_64-unknown-linux-musl

Security Best Practices in Rust

While Rust’s safety guarantees are a strong foundation, additional best practices can further enhance application security:

  • Minimize unsafe blocks: Limit the use of unsafe code to avoid memory vulnerabilities.
  • Use password hashing for sensitive data: Store passwords using Argon2, bcrypt, or scrypt, not as plaintext.
  • Leverage strong typing and lifetimes: Rust’s type system prevents common errors by ensuring proper data handling.
  • Employ secure libraries: Use libraries like ring, rustls, and argon2 rather than implementing cryptographic functions, as custom cryptography is challenging to secure.

Security Audits and Code Analysis

Rust’s ecosystem includes tools for static analysis and security auditing, such as cargo-audit, which checks dependencies for known vulnerabilities.

cargo install cargo-audit
cargo audit

cargo-audit is especially useful for detecting security issues in third-party libraries.

Secure Memory Management

Rust’s zero-cost abstractions ensure safety without sacrificing performance, which is critical for secure memory handling. Libraries like secrecy help secure data in memory, preventing leaks and ensuring sensitive data is cleared when no longer needed.

Using secrecy for Sensitive Data

The secrecy crate provides secure wrappers around sensitive data types, ensuring they’re wiped from memory when dropped.

use secrecy::{Secret, ExposeSecret};

fn main() {
    let password = Secret::new(String::from("super_secret_password"));
    println!("Password: {}", password.expose_secret());
}

secrecy is useful for managing in-memory secrets, ensuring sensitive data is not accidentally leaked.

Summary

Rust’s security-focused libraries, memory safety guarantees, and secure-by-default principles make it ideal for developing cryptographic applications and secure systems. With tools for encryption, password hashing, TLS, cross-compilation, and secure memory handling, Rust provides a strong foundation for building secure, high-performance applications.