The old style Airport Expresses are amazing and have been around a long time, they are great for extending your home network and also great for setting up impromptu networks when travelling.
However, as of Mountain Lion the Airport Utility (which is now pretty dumbed down) doesn’t support them which is kind of lame – I guess when the devices once had an expected life of 18 months that isn’t much of a surprise.
There is however still a solution, AirPort Utility 5.6 for Mac OS X Lion, which of course you can’t install on Mountain Lion until you find these excellent instructions.
From frustrated to happy in the space of 5 minutes, now I can get this Tiger running Mac Mini from 2006 back online alongside the new Mac Book Air that replaces it.
Now I just need to see if I can set the second one up for my in-laws too so they can get iPlayer on the TV
Every now and then I find myself needing to quickly analyse a set of access_log files to see who the most common visitors are so that I can decide if there are any abusers I should be blocking or poorly configured services running somewhere that I can try to get fixed. I can never remember the quickest was to do this so I decided to write down the “one liner” that I cobbled together this time so I can hopefully find it next time and not have to reinvent the wheel again.
Here is the one-liner I used to find the top IPs this time:
sed -e 's/\([0-9]\+\.[0-9]\+\.[0-9]\+\.[0-9]\+\).*$/\1/' -e t -e d access.log | sort | uniq -c | egrep -v "\s([0-9]|[0-9][0-9]|[0-9][0-9][0-9]) "
Splitting this out we have:
- A call to <code>sed</code> to extract all the IP Addresses from the access_log file
- A call to <code>sort</code> to sort the list of IPs
- A call to <code>uniq</code> to create a list of unique IPs with counts
- A call to
egrep to filter the unique list down to IPs we at least 1000 appearances – this will need tuning depending on the volume of requests / time period the file covers.
Last night I did a short presentation on Scaling WordPress using WordPress.com as an example giving an overview of the solutions we use for different scaling issues as well as highlighting some solutions that you can try out on smaller sites:
As I sit here blaming code to find the original source of a line of code I’m beginning to think that
svn needs a new improved version of
spelunk it would work something like this:
$ svn help spelunk
spelunk (curse, showup, show): Output the content of specified files or URLs with the original revision and author information in-line ignoring white space changes and following movement of code between files.
Yes – I know I am dreaming