Exploring NAT Traversal and Tunnels with Inlets and Inlets Pro
Inlets! What is it?
inlets is a reverse proxy and service tunnel written in Go.
This is the section from the inlet’s github README
inlets combines a reverse proxy and websocket tunnels to expose your internal and development endpoints to the public Internet via an exit-node. An exit-node may be a 5-10 USD VPS or any other computer with an IPv4 IP address.
Why do we need this project? Similar tools such as ngrok or Argo Tunnel from Cloudflare are closed-source, have limits built-in, can work out expensive, and have limited support for arm/arm64. Ngrok is also often banned by corporate firewall policies meaning it can be unusable. Other open-source tunnel tools are designed to only set up a single static tunnel. inlets aims to dynamically bind and discover your local services to DNS entries with automated TLS certificates to a public IP address over a websocket tunnel.
When combined with SSL - inlets can be used with any corporate HTTP proxy which supports CONNECT.
This means you can use it to create inbound network tunnels to computers that don’t have a public IP, are behind firewalls or get assigned new IPs frequently. Things like your laptop when you move from the office to a coffee shop. While writing this post I have probably been on 5 different networks, but still been able to connect to my laptop’s k3d cluster using the same public IP address.
It’s on the landscape in the section. It’s cloud native! inlets has a place on the Cloud Native Computing Foundation (CNCF)’s technology landscape. It’s in the Service Proxy category along with MetalLB, Traefik, HAProxy and Nginx.
There are plenty of freemium services out there that will give you inbound network access to processes that don’t have inbound connectivity, but most of these services are in control of your data, rate limit and are closed source. Inlets gives you back control of your data, privacy and lets you run the service in the way you want.
We have already mentioned a few services like Ngrok and Argo Tunnel, but they are closed source and have severe service restrictions, plus you are not fully in control of the service.
With cloud VMs starting at $5 with heaps of inbound data transfer there’s no need to pay 10s of dollars a month for someone else to run a limited service for you.
How it works
Inlets creates a tunnel from your local process to a public cloud instance (or anything that can accept inbound traffic). I have been using digital ocean’s smallest VMs as my exit nodes. However, you can set up inlets using anything that can be accessed from both the internet, and the application or process you want to expose.
There is a handy CLI for managing your inlets services called inletsctl. I will be using this cli throughout the post as it simplifies the process of getting things set up and configured.
_ _ _ _ _ (_)_ __ | | ___| |_ ___ ___| |_| | | | '_ \| |/ _ \ __/ __|/ __| __| | | | | | | | __/ |_\__ \ (__| |_| | |_|_| |_|_|\___|\__|___/\___|\__|_| inletsctl can create exit nodes for you on your preferred cloud provider so that you can run a single command and then connect with your inlets client. See also: inlets-operator for Kubernetes and inlets-pro for TCP tunnelling. Usage: inletsctl [flags] inletsctl [command] Available Commands: create Create an exit node delete Delete an exit node help Help about any command kfwd Forward a Kubernetes service to the local machine version Display the clients version information. Flags: -h, --help help for inletsctl Use "inletsctl [command] --help" for more information about a command.
The create command automates the provisioning, setup and configuration of the publicly accessible exit-node using cloud instances.
Currently supported providers:
- Digital Ocean
- Google Cloud
You will need your own account and will incur costs with whichever provider you use if you are following along.
There are some great blog posts already about setting up inlets without the CLI
- HTTPS for your local endpoints with inlets and Caddy - Alex Ellis
- Setting up an EC2 Instance as an Inlets Exit Node - Matt Bacchi
Therefore I won’t be focusing on this, instead we will focus on the why and the what of inlets.
Before you get started
It’s probably best to say that if you are using a corporate network, or any network you aren’t responsible for, you should ensure you have permission to start routing inbound internet traffic. It goes without saying that this exposes the network to extra risk and could drastically increase the network attack surface.
I’m using Kubernetes and can’t forward every deployment, that would take ages…
Not to worry, the inlets-operator has your back.
The operator automates the creation of an inlets exit-node on public cloud, and runs the client as a Pod inside your cluster. Your Kubernetes Service will be updated with the public IP of the exit-node and you can start receiving incoming traffic immediately.
I have been writing a separate post on the inlets operator. It has a much more specific usecase if using Kubernetes somewhere that you don’t have inbound network traffic. Like k3s/k3d/minikube/microk8s/Docker Desktop/KinD.
It’s worth exploring if you need to expose your local kubernetes services, for example to give a demo or collaborate.
People LOVE reverse proxies
My most viewed tweet so far is this
I’m not sure I fully get why people seem to love tunneling network traffic into networks where it’s either specifically disallowed (Corporate networks).
Why Inlets Pro?
Inlets Pro is a commercial extension to inlets, you can try it out and contact OpenFaaS for more information
According to the github REAMDE inlets-pro is an L4 TCP load-balancer.
For those of us that can’t remember the entire network stack and what each layer does or is (humans?),
this roughly translates as
deliveres the correct data to the correct application process over the network.
So putting that in the context of inlets-pro, it forwards TCP traffic over an encrypted websocket secured with TLS to your application process. Cool!
These are some of the features
- Support for any TCP protocol
- Automatic TLS encryption for tunnel and control-port
- Pass-through L4 proxy
- Automatic port detection, as announced by client
- systemd support and automatic retries
- Kubernetes compatible
If you just want to expose a service for http traffic you can use inlets. I use standard inlets when proxying my draft blog posts, which are served from my laptop, out to the internet so my friends and family can proofread them. (I’m REALLY bad at spelling)
But for this post we will use both standard and pro features to show off as many use cases as possible.
Enough talk - time for action
We have already briefly looked at
inletsctl, but we didnt discuss installation or setup.
All of the commands shown are for linux and will probably work on Mac too. If you are using Windows then you should probably use git bash
Download and install inlets, inlets pro and inletsctl
# Install Inletsctl to /usr/local/bin curl -sLSf https://inletsctl.inlets.dev | sudo sh # Install Inlets sudo inletsctl download # Install Inlets Pro sudo inletsctl download --pro
I’m using Digital Ocean for my exit nodes, you need to grab your Digital Ocean API Key if you are following along.
I have saved mine in a file called
do-token in my home directory.
If you want to use one of the other providers you can check the examples in the inletsctl github repo.
Then we can create an exit node
inletsctl create --access-token-file ~/do-token
You should have some similar output to above, this is showing that we created a node and it’s ready to be configured to forward traffic. It’s even helpfully given us the commands we need to run to get something configured.
If you have docker installed you can do a quick test to see if it’s working.
docker run --name nginx-test -d -p 8080:80 --rm nginx:latest export UPSTREAM=http://127.0.0.1:8080 inlets client --remote "ws://184.108.40.206:8080" \ --token "fzjFpqorbwS4Zi06CQNxeSNOorlEk2z4qteh6ybCTHmOwvPcHL25hunDrHM5K5zw" \ --upstream $UPSTREAM
You should have some output in your terminal that shows inlets client connecting to your exit server. You can now navigate to the public IP of your exit node. If you see the nginx splash page then we are all set up!
If that worked you can stop the docker container
docker stop nginx-test
Now we have validated our setup we can get into some real use cases!
Broadly my use cases are split into HTTP traffic, and non HTTP traffic. The example commands to forward these two categories of traffic are the same, so I will show some example uses and then the commands needed to get inlets or inlets-pro working.
HTTP based services
While developing OpenFaaS Cloud we rely on webhooks to let us know when new code has been pushed, and therefore when to run new builds.
This is easy with inlets, we just forward the webhooks through an inlets exit-node to our local machine, this means we can build our services locally without having to push containers up to registries and pull them back down in a remote kubernetes cluster. This saves us loads of time and bandwidth.
Exposing temporary services
Let’s suppose you are working remotely on a set of microservices, your colleague wants to test their new front-end against your backend service, this means committing, raising a PR and getting code into a development or staging environment to properly test. Alternatively they could pull your code, build and deploy that service themselves.
Problem is, both of these are a hassle. We like things to be easy! You get some free user testing, potentially catch a bug before your code even makes it into PR, your colleague gets a service provided by the person building it, without having to do anything. Sounds good.
All you need to do is expose your service using inlets. The commands to do this are later in the section.
It’s always cool to give access to the demo services and websites you are making to the “users”, this could be the product owner, client or your business partner.
Without using something like inlets available you would need to deploy your application up to a machine that had a public IP, configure the service and then remember to keep that up-to-date. You would also need to go off to that machine to get logs. Having this instance running costs you money and takes away valuable time during the early product development stages.
Using inlets we can expose a local service to the internet in less than a minute, and when we are done we can tear everything down again, meaning we can create on-demand tunnels, only pay for what we use and spend more time where it matters, creating value for our customers.
I can deploy a local copy of my blog using docker-compose, that updates when files change and allows me to see what my changes will look like without publishing to the rest of the world.
Let’s look at the last example above, my personal blog in development mode.
I can expose my service to
http://127.0.0.1:4000 by running
docker-compose up. This stands up a container that
publishes to port 4000 on my local machine.
Using the exit-node we created earlier we can set up an inbound network route to my blog like this
export UPSTREAM=http://127.0.0.1:4000 inlets client --remote "ws://220.127.116.11:8080" \ --token "fzjFpqorbwS4Zi06CQNxeSNOorlEk2z4qteh6ybCTHmOwvPcHL25hunDrHM5K5zw" \ --upstream $UPSTREAM
My local service will be accessible at the remote ip. That’s it, all that’s needed.
You can substitute the port on the
UPSTREAM environment variable if your service is using a different port,
5000 for example:
Then we just run the
inlets command again.
Non-HTTP based services (inlets pro)
I didn’t know if I should have put Home Lab under HTTP services, but I decided it probably lives under the inlets-pro section, I personally use mainly HTTP, but do have an SSH Tunnel open from time-to-time to allow configuration of the resources from outside my home network.
You might be collaborating with someone who doesn’t have their own resources, or computer with particular chipset. So it’s useful to be able to open these up to trusted parties via the internet.
Here’s an example from the inlets author
You could open a reverse tunnel to configure, update and change the software running on IoT Devices no-matter where they are, provided they have outbound network connectivity.
This could be a really good way to deliver updates to IoT devices.
Sometimes you need to get access to resources in private networks. This could be a database cluster or your kubernetes nodes for example. A common pattern is to have “SSH Bastions”. These are machines that live within the network but have an inbound network route. This allows you to ssh onto these machines and then jump into internal parts of your network.
Usually access to these machines is usually controlled by IP Whitelisting, using SSH Keys and .
We could add an extra layer of security into the mix here by not having any inbound network access to the machine. This limits the discoverability of the service. If you have ever run SSH on a public IP you will be inundated by bots trying weak username/password combinations.
This would look like this:
An example, SSH
So to set up TCP forwarding we can use
inletsctl. I will be running through a simple SSH example,
but this could be applied to any of the non-HTTP examples in the same way.
I’m going to run the client on one of my Raspberry Pi computers with ssh enabled using key pairs only (here is a Digital Ocean post on setting this up)
Once we are logged into our Pi on our local network we can set up the inlets-pro client to enable ssh access from the IP of our exit node. We also need to change the default SSH port away from 22, as our exit-node will have ssh mapped to this port and we won’t be able to use it.
Here is an article about changing the default ssh port I used 2222, and will refer to that from now on.
# Set this to the IP of the machine you want to expose on the web export RPI_IP=192.168.0.99 inletsctl create --access-token-file ~/do-token --remote-tcp $RPI_IP
This gives you a command that looks like this to run on our client to forward the traffic
export TCP_PORTS="2222" export LICENSE="YOUR INLETS_PRO_LISENCE" inlets-pro client --connect "wss://18.104.22.168:8123/connect" \ --token "a-long-random-string" \ --license "$LICENSE" \ --tcp-ports $TCP_PORTS
We need to forward port
2222, so set this in the
TCP_PORTS variable. Don’t forget to set your inlets-pro licence key
The last step is starting the client on our Raspberry pi. SSH on there from your computer and run the command!
If we have set up SSH correctly, and everything works as expected we should be able to ssh onto our Pi using the public IP
ssh -p 2222 email@example.com
This example should work for any service that uses TCP. You will just need to edit the
TCP_PORTS environment variable
and re-run the client command to forward different ports.
Have a go yourself
Check out each of the tools we have looked at on github
- inlets - the HTTP tunnel
- inlets-pro - the TCP tunnel
- inletsctl - The cli tool we use to speed up provisioning servers
I’m sure there are loads of other things we could do with this. I would love to hear from you if you have any more great use cases. You can tweet me @alistair_hey or use linkedin, github or email (links should be at the bottom of the page)