Skip to content

crookedstorm/fibsrv

Repository files navigation

FibSrv - A little Fibonacci number generator

This is a little coding exercise that aims to try to squeeze speed out of basic golang where possible to return the nth Fibonacci number in response to GET http requests at $uri/api/fibonacci/:number where :number is the nth position in the sequence you'd like to see.

If you enjoy watching metrics, you can also view prometheus metrics (or ideally scrape them with prometheus) at $uri/metrics. The server will log to stdout.

When running locally, this defaults to a URI of localhost:3000.

Building

If you like to try things without docker, and presuming you have a recent version of Go that includes support for modules, you should be able to just run: make build at the root of this repository. This will produce a binary at bin/fibsrv.

Running tests

The really easy way here is make test. If you want to try out the benchmarks, you can run cd pkg/fibber && go test -bench=. to see the wild difference between calculating big.Ints with the go code vs just doing a simple lookup in an array.

Docker Compose

If you want things to be a bit tidier, you can run make up and you'll get a full deployment with nginx out front, 6 fibsrv workers behind it, prometheus monitoring (not sending alerts to anywhere in particular) and likely cadvisor (works on a family member's intel machine so probably yours as well).

I recommend you try this method.

When done, run make down if you want to type a little less than docker-compose down.

Docker

You can build a docker container with make docker_build. This will tag the image to make it fairly easy to run with something like docker run -e GIN_MODE=release -p 3000:3000 docker.io/crookedstorm/fibsrv:0.0.1 or you can leave off the environment variable for release mode if you want the default debug mode. However, you really probably should use...

The Wish List

This exercise is time-boxed and done in spare time in general. As a result this is going to be a pale reflection of a production system. Here are some ways I would ramp this up for production if the world really was in need of quickly obtained Fibonacci numbers from the web.

I'm deliberately not putting backups on this list becuase we don't have a lot of state to worry about. I'm going to blithely trust that my git server will keep my code safe enough and the various services will still be availble from Dockerhub tomorrow. For some applications, this isn't always a perfect assumption, but I'll go with it here. If I was that worried about that, all the code would be forked, the containers would be cached in a registry, and I would have my bunker stocked.

A Helm Chart

It would be quicker and more flexible for Kubernetes deployment to chart things out, add a horizontal pod autoscaler, be ready with the ingress for whatever controller our cluster is using and take care of all those little concerns.

It probably wouldn't be too hard to squeeze in an optional TLS cert from certmanager/Let's Encrypt that's actually valid, all that the docker compose does and then some. I inherited the care and feeding of the helm setup for PAWS, which is mostly using a helm chart to customize and bring together other helm charts. That sort of setup would be lovely for setting things up given a little time and testing. This also implies that there's a k8s cluster to plug into.

External Monitoring

The telescope was invented for more reasons than viewing the sky. It pays to be able to tell if the entire system is down, not just one pod. You cannot get an alert from inside a network that cannot communicate, for instance.

Remote Logging

You really shouldn't dockerize things without sending the logs somewhere. It just makes things harder.

Staging Environment

SLOs are tricky to develop at the best of times, but if you are testing performance on dramatically different hardware than you might be deploying on, it's just guessing.

I'm developing this on an ARM64 Mac. There isn't much chance of that being the deployment platform. Given a bit of time, I might want to try it out on my staging environment and then beat on it a while to decide just what acceptable looks like.

Histogram Buckets in Prometheus

I used a quick open source solution for exporting prometheus stats instead of concocting my own middleware. That's often a great time saver, but if we are working on performance SLOs, quantiles would be very nice. That would take me a little more time.

Grafana (or similar)

Prometheus is nice, but it really doesn't "give" you much without some dashboards made while you had some time to think about it. Dashboards and visualizations are a gift to your future self that allow you to see things you didn't know you were looking for and find the patterns that actually help you decide if that SLA you agreed to was a bad idea.

Beating on it

The benchmarks and tests in the go code looked promising, but I know that I'd done nothing to stop golang from just running out of threads as it takes load

When I cranked up rakyll/hey a little (to 250 workers and 600 requests), I thought I had what I was after (even using the precalculated array of numbers):

 hey -n 600 -c 250 http://localhost:3000/api/fibonacci/20

Summary:
  Total:	0.1168 secs
  Slowest:	0.1013 secs
  Fastest:	0.0015 secs
  Average:	0.0549 secs
  Requests/sec:	4282.1769

  Total data:	1392 bytes
  Size/request:	4 bytes

Response time histogram:
  0.001 [1]	|■
  0.011 [44]	|■■■■■■■■■■■■■■■■■■■■■■■■■■
  0.021 [55]	|■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■
  0.031 [3]	|■■
  0.041 [34]	|■■■■■■■■■■■■■■■■■■■■
  0.051 [5]	|■■■
  0.061 [20]	|■■■■■■■■■■■■
  0.071 [20]	|■■■■■■■■■■■■
  0.081 [68]	|■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■
  0.091 [60]	|■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■
  0.101 [38]	|■■■■■■■■■■■■■■■■■■■■■■


Latency distribution:
  10% in 0.0087 secs
  25% in 0.0132 secs
  50% in 0.0698 secs
  75% in 0.0870 secs
  90% in 0.0920 secs
  95% in 0.0970 secs
  99% in 0.1004 secs

Details (average, fastest, slowest):
  DNS+dialup:	0.0078 secs, 0.0015 secs, 0.1013 secs
  DNS-lookup:	0.0034 secs, 0.0000 secs, 0.0094 secs
  req write:	0.0002 secs, 0.0000 secs, 0.0029 secs
  resp wait:	0.0393 secs, 0.0014 secs, 0.0803 secs
  resp read:	0.0000 secs, 0.0000 secs, 0.0002 secs

Status code distribution:
  [200]	348 responses

Error distribution:
  [1]	Get "http://localhost:3000/api/fibonacci/20": read tcp [::1]:57091->[::1]:3000: read: connection reset by peer
  [1]	Get "http://localhost:3000/api/fibonacci/20": read tcp [::1]:57092->[::1]:3000: read: connection reset by peer
  [1]	Get "http://localhost:3000/api/fibonacci/20": read tcp [::1]:57093->[::1]:3000: read: connection reset by peer
  [1]	Get "http://localhost:3000/api/fibonacci/20": read tcp [::1]:57095->[::1]:3000: read: connection reset by peer
<cut off for brevity>

Unfortunately, after scaling up my app to 6 servers with an nginx server in front using roundrobin docker compose DNS to kind of load balance the whole deal, it still does that becuase the error is actually from the client (hey). That was annoying. Apparently on my mac, rakyll/hey will (even with ulimit increased) eventually have spasms if I go much over 100 concurrent workers. That made me sad. Anyway, I'm including my final version with nginx plus 6 workers and hey attacking with 100 concurrent workers and the calculated big numers version (which is totally based on the example offered in the golang source code for the big module, I should add):

┌─[brooke][callisto][±][main {1} U:2 ✗][~/src/noodling/do-assignment]
└─▪ hey -n 6600 -c 100 http://localhost:4000/api/fibonacci/180

Summary:
  Total:	1.5451 secs
  Slowest:	0.1633 secs
  Fastest:	0.0008 secs
  Average:	0.0202 secs
  Requests/sec:	4271.6188

  Total data:	250800 bytes
  Size/request:	38 bytes

Response time histogram:
  0.001 [1]	|
  0.017 [3816]	|■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■■
  0.033 [1627]	|■■■■■■■■■■■■■■■■■
  0.050 [750]	|■■■■■■■■
  0.066 [222]	|■■
  0.082 [91]	|■
  0.098 [31]	|
  0.115 [18]	|
  0.131 [10]	|
  0.147 [6]	|
  0.163 [28]	|


Latency distribution:
  10% in 0.0043 secs
  25% in 0.0072 secs
  50% in 0.0141 secs
  75% in 0.0284 secs
  90% in 0.0423 secs
  95% in 0.0531 secs
  99% in 0.0929 secs

Details (average, fastest, slowest):
  DNS+dialup:	0.0001 secs, 0.0008 secs, 0.1633 secs
  DNS-lookup:	0.0001 secs, 0.0000 secs, 0.0045 secs
  req write:	0.0000 secs, 0.0000 secs, 0.0014 secs
  resp wait:	0.0200 secs, 0.0008 secs, 0.1562 secs
  resp read:	0.0001 secs, 0.0000 secs, 0.0102 secs

Status code distribution:
  [200]	6600 responses

That's pretty fast, I think. It's not doing much (no databases or queues to slow things down), and it's probably way more parallel than it needs right now. I could hammer on it more using different tools if I took more time.

About

Totally an interview question toy

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published