What level of hardware resources are you using to run your blog/wiki/side project/work site, and what level of traffic does it normally sustain?
I’m trying to gather a little data based on this thread. I’d like to either publish said data in some kind of wiki, or find a community on the web that tracks these kinds of things.
VMs are hosted by brightbox.com (mostly because I used to work there and the service is good/UK based which suits me.)
We’ve started up a repo to collect suggestions!
[Comment removed by author]
Off-topic: What are you running for your git there? I’m looking into GitLab on a Raspberry Pi 3, which looks thoroughly supported and simple.
I looked at code.devlinzed.com; is that gitweb interface?
[Comment removed by author]
Thank you. :-)
I ran a not very complicated PHP website on a 128MB VPS a few years ago. It got around 4K hits per day usually, peaked at 20k.
I run several sites on wholly unremarkable hardware. As in, rails on the cheapest I could possibly find. As long as you have a cache in front (apache mod_cache, nginx or varnish) and your backend produces a Cache-Control header, your software will handle ~all load peaks. High traffic level involves repeated requests and the cache takes care of precisely those.
So my answer is: Just add a cache, make sure your pages are cache-friendly, then stop worrying about this. Worry about the content instead.
I have a Django site (some old version, I can’t remember off the top of my head) running on nginx + postgres on a 1GB Rackspace VPS that has withstood some pretty impressive traffic spikes in the past (>100,000 hits in a day) and routinely gets ~25,000 hits/month.
Are you not worried about security on an old version of Django??
Not really, there’s no personal data, and if it gets hacked that will give me the motivation to rewrite it.