I’ve been playing around with Docker a lot recently, and one thing that seems to be lacking is a way to update all of your downloaded images in an “apt-get upgrade” fashion. I know this isn’t usually needed in most cases, but in a home use case for Docker most of the time, you can upgrade everything in one go and not worry about it. So I wrote some quick one-line scripts in Bash and PowerShell that can be used to update all your docker images. Continue reading Updating all your Docker images in one swoop.
Ok, so it’s been awhile since I put up the last part. So long that I actually ended up looking at some other hypervisor options, and still ended up going with Proxmox. I’ve also made it slightly more robust by running two host machines in a cluster configuration. I not running many virtual machines right now, but the ones that I do run are mostly OpenVZ containers from TurnkeyLinux.org, which Proxmox has a built-in integration to download templates straight from them. There are also other OpenVZ templates from Proxmox for a few standard linux server OSes, but the Turnkey Linux ones seem better at being kept up to date. Below the fold is a short list of what I’m currently running and a few that I plan on setting up at some point. Continue reading Running a Home Virtual Server – Part 2: Contemplating Containers
So I’m working on a project at work to evaluate and implement a source control system to replace a fairly non-existent use of Visual SourceSafe. It has been a while since I’ve looked really in depth at any version control software and I have to say that I’m impressed with the advancements around them. Not necessarily in the version control systems specifically, but in a lot of the auxiliary functionality that they can offer either natively or as with easy add-on. Things like automated builds and continuous integration are things that weren’t even on the map last time I looked at this in any depth.
At my work we decided to evaluate three types of version control; Team Foundation Server (TFS), Subversion, and Git. I’m most familiar with Subversion since that’s what I’ve used in the past and TFS is a fairly obvious choice for us since we are mostly a Microsoft shop with the tools we use. Git however was a bit of a research project for me and I have to say that we fairly quickly decided that it wasn’t for us. Don’t get me wrong it has it’s use cases where it shines, particularly in development environments where the development team is widely dispersed geographically and not necessarily always network connected. In these cases Git seems really ideal which is probably why it has taken off and is so popular in open source projects, but this is completely different from our primary development environment and the downsides outweigh these benefits.
I don’t know which way we’ll end up going at my job. We’re wanting something simple that will be easy to use and we’re just implementing source control as part of the project. This is why I’m starting to think that we might go with Subversion because TFS can be such a bear to administrate and after getting them installed and starting to set them up, I can understand.
So everything has been pretty busy recently and here’s a quick update for some projects. I had some hardware problems with my home virtual server project. I think it came down to a motherboard that went bad, so I changed out the internals and then had some “less than graceful” shutdowns. I think this might have caused some trouble with terabyte drive, but interestingly the RAID 10 setup seems to be fine. In switching out the hardware I did have to downgrade as well, so I bought another Core 2 Duo from my university’s surplus sale recently and upgraded it back to original specs. I’ve also thought that I would dedicate the mass storage into a NAS build from another surplus computer. I’ll have some more about each of these in the near future.
In this last post I talked about picking up a computer at my university’s surplus sale to use as a replacement server in my home setup. The plan is to use this machine to run several virtual machines on, mostly to play around on but also as my home server. I didn’t go into to much detail in that first post, so I’m going to break this down into multiple posts that can go into more detail. I’m not sure how many parts there will be as of now, but I’ll be sure to try and link all the posts together so they should be easy to find. First up is a discussion about the hardware and my final decision about what virtual environment I decided to go with.
There are pros and cons to working on any kinds of projects, and a setting up a home virtual server is one of those projects that you really need to weigh each and see if you’re up for it. Having started well down the path of implementing one in my home, I think for me it’ll be worth it in the long term going this route. I’ll tell you some of the benefits and drawback I think I’ll have as well as some of the requirements I started out with in my head.
I came to a sad realization about a week ago now. My university will occasionally have a surplus sale about once a quarter and at least once a semester. They usually sell stuff like desks, chairs, and old couches from the dorms (disgusting). They also use this way to unload computers that get replaced and can’t find a home somewhere else in the school. I’m lucky because they sell computers individually (monitors separate), where a lot of other schools I’ve seen will bundle a pallet and sell them at surplus auctions. In order to sell them individually they have to sell them really cheap ($2 per desktop, $4 per LCD monitor, $10 per laptop) and this will draw quite the crowd. People will camp out the night before in order to be the first in line to get some computers/laptops, it’s a bit like watching footage of Black Friday shoppers.