If your development setup is anything like mine, you’ll like to put all of your applications into their own containers so that they’re isolated from each other. This also gives me a little added guarantee that all of an application’s dependencies are wrapped up nicely before moving the code between environments.
Sometimes, debugging can be a little awkward if this is how you run. In today’s post, I’ll take you through debugging your node apps inside of a container.
Execution
The execution environment is quite simple. We’ll assume that a bash script allows us to start a container which holds our application, and injects any instruction to its console:
Our application gets mounted to /usr/src/app inside the container
Allowing inspection
Now we need to tell our node process that we want to inspect the process, and allow debugging. This is as simple as using the --inspect switch with your node or in my case nodemon invocations. Here is my debug run script inside of my package.json:
This starts execution, mounting the debug port on 9229 (to align with our docker invocation); it’s also allowing connections from any remote computer to perform debugging. Handy.
Start debugging
Once you’ve issued ./run npm run debug at the console, you’re ready to start debugging.
I use WebStorm for some projects, vim for others; and sometimes will use Chrome Dev Tools with chrome://inspect to be able to see debugging information on screen.
Hope this helps you keep everything isolated; but integrated enough to debug!
When you’re developing an AWS Lambda, sometimes you’re going to need to install binary package dependencies. Today’s article will take you through the construction of a project that can be deployed into AWS Lambda including your binary dependencies.
Structure
The whole idea here is based on AWS Lambda using Docker to facilite package, deployment, and execution of your function. The standard python:3.6 image available in the standard library is compatible with what we’ll end up deploying.
The structure of your project should have a requirements.txt file holding your dependencies, a standard Dockerfile and of course, your code.
Any depeendencies are listed out by the requirements.txt file.
Docker
We can now bundle our application up, so that it can be used by AWS Lambda.
FROM python:3.6
RUN apt-get update && apt-get install -y zip
WORKDIR /lambda
# add the requirements and perform any installations
ADD requirements.txt /tmp
RUN pip install --quiet -t /lambda -r /tmp/requirements.txt && \
find /lambda -type d | xargs chmod ugo+rx && \
find /lambda -type f | xargs chmod ugo+r
# the application source code is added to the container
ADD src/ /lambda/
RUN find /lambda -type d | xargs chmod ugo+rx && \
find /lambda -type f | xargs chmod ugo+r
# pre-compilation into the container
RUN python -m compileall -q /lambda
RUN zip --quiet -9r /lambda.zip .
FROM scratch
COPY --from=0 /lambda.zip /
The docker container is then built with the following:
The open standard of JSON Web Tokens allows parties to securely transfer claim information. Furthermore, signed tokens allow for verification of the token itself.
One of the most common use-cases of the JWT is for authorization. Once a user’s secrets have been verified, they are issued a JWT containing claim information. This token is then attached to further requests that allows server applications to assert user’s access to services, resources, etc.
JWTs can also be used for adhoc information transfer. A token’s ability to be signed is an important characteristic providing information verifiability between parties.
Structure
A JWT is divided into three primary components:
Header
Payload
Signature
The header contains information about the type of token; this article assumes that JWT is the type and the hashing algorithm. An example header might look something like this:
{
"alg": "RSA",
"typ": "JWT"
}
The payload part is expected to have some standard attribute values. These values can fit into the following categories:
Type
Description
Registered claims
“iss” Issuer, “sub” Subject, “aud” Audience, “exp” Expiration Time, “nbf” Not before, “iat” Issued at, “jti” JWT ID
Public claims
Freely definable names that should be made collision resistant.
Private claims
Private claims don’t need to be collision resistant. Use these with caution.
A simple payload for a user might look something like this:
Finally, the last piece of the JWT is the signature. The signature of the message depends on the hashing algorithm that you’ve selected in your header.
The calculation is going to look something like this:
Configure authentication using a home-folder credential file called .passwd-s3fs. This file expects data in the format of IDENTITY:CREDENTIAL. You can easily create one of these with the following:
Like it or hate it, you’ll find systemd on any modern linux system these days. This management subsystem is the newer replacement for the older init scripts-based systems.
In today’s article, I’m going to show you how to create a daemon that will sit under the management of systemd.
.service file
A service file instructs systemd how to interact with the deployed application. It’s referred to as the service unit configuration. The following is an example:
[Unit]
Description=My Service
After=network.target
[Service]
ExecStart=/usr/bin/my-service
Restart=on-failure
[Install]
WantedBy=multi-user.target
Installation
This file, once you’ve created it gets deployed to systemd with a cp. You also need to notify systemd.