

Linkerd Traffic Split
You can shift a portion of incoming traffic to different destination services with linkerd’s traffic split functionality. This feature becomes handy when you want to deploy a newer version of your service by increasing traffic slowly while making sure everything is alright. (Canary Release, Blue Green Deployment etc.)
Prerequisites
To follow this lab you’ll need linkerd installed on your cluster. Follow this guide if you didn’t already.
In this lab we’ll use two docker images (oktano/sample-pizza-server-app:v2, oktano/sample-pizza-client-app:v1). Pizza client generates traffic on pizza server. Pizza server takes two environment variables so we can adjust success rate and latency of it. For initial state set success rate to 10 (which means 100%) and latency to 100ms.
You can run commands below to get to initial state.
kubectl create ns pizza-app
# Add linkerd.io/inject annotation for linkerd's automatic proxy injection
kubectl annotate ns pizza-app linkerd.io/inject=enabled
# Deploy pizza server application
kubectl run pizza-server-v2 --image oktano/sample-pizza-server-app:v2 --env SUCCESS_RATE=10 --env MAX_SLEEP_DURATION=100 -n pizza-app
# Expose pizza server application
kubectl expose deployment pizza-server-v2 -n pizza-app --port 5000 --target-port 5000
# Deploy pizza client application
kubectl run pizza-client --image oktano/sample-pizza-client-app:v1 --env CALL_CENTER_URL=http://pizza-server-v2:5000 -n pizza-app
Wait for a little bit and verify that your deployment’s success rate is 100%.
linkerd -n pizza-app stat deploy
# I removed unrelevant columns from output
NAME SUCCESS
pizza-client -
pizza-server-v2 100.00%
Lets deploy our ‘newer version’ of pizza-server. We’ll use same image with different success rate and shift some portion of traffic to it. Deploy and expose ‘v3’ with command down below. You can set success rate and max sleep duration to whatever you want.
kubectl run pizza-server-v3 --image oktano/sample-pizza-server-app:v2 --env SUCCESS_RATE=0 --env MAX_SLEEP_DURATION=100 -n pizza-app
kubectl expose deployment pizza-server-v3 -n pizza-app --port 5000 --target-port 5000
Now shift some traffic to v3. Create a TrafficSplit configuration and linkerd will take care of the rest. Again, you can play with these values. I’ll shift 10% of traffic to v3 for start.
cat <<EOF | kubectl apply -f -
apiVersion: split.smi-spec.io/v1alpha1
kind: TrafficSplit
metadata:
name: traffic-split-v3
namespace: pizza-app
spec:
service: pizza-server-v2
backends:
- service: pizza-server-v2
weight: 900m
- service: pizza-server-v3
weight: 100m
EOF
After that linkerd will slowly start to shift traffic to pizza-server-v3 service. You can see it in dashboard. Check the RPS values.

Linkerd shifting traffic to new version
pizza-server-v2 deployment’s success rate is still 100%. Because it is a seperate deployment from pizza-server-v3. To see overall success rate from pizza-client deployment’s point of view run:
linkerd -n pizza-app routes deploy/pizza-client --to service/pizza-server-v2
ROUTE SERVICE SUCCESS ...
[DEFAULT] pizza-server-v2 89.17% ...
Cleanup
To remove everything from this lab use kubectl delete ns pizza-app command.