Category: "Ubuntu"

Pages: 2 4 5


  10:30:00 am, by The Dreamer   , 2100 words  
Categories: Hardware, Software, Computer, Storage, Ubuntu, FreeBSD, VirtualBox

Orac is looking strangely bare, with Zen taking over.

For a long time, I've been running a 6 drive RAID 10 array of Hitachi 5K3000 2TB drives in Orac for backuppc. This configuration got me at 5.4TB array, and somewhat better performance than when I tried a RAID6 configuration. But, eventually, I kept running out of space and the price of harddrives went up so expanding the array over time didn't happen as I had hoped. Being RAID10, the options were other concat another array, either 2 in RAID1 or 4 in RAID10 or 4 as 2 RAID1....using volume manager. Or maybe see if RAID10 would deal with having all 6 drives upgraded to 3TB, though hadn't considered the transition of 512 to 4k and how it would cope with that.

Though I did, eventually find out when I upgraded a 1.5TB RAID1 set to become a 2TB RAID1 set....going from ST31500341AS to ST2000DL003, where I contributed my experience here: Though it first started because one of the ST31500341AS drives had failed.

Before the failure of one of the 1.5TB drives in the above mentioned RAID1 set, I had 4 ST31500341AS in a RAID5 on old-Zen. It had been done in under RR622, under Windows, and NTFS partitions, etc. I had tried copying the data at various times, not really having anywhere that would hold the data elsewhere...but wanting to get it over to FreeBSD for recovery. While I got the rr622 driver working, and it saw that I had a single array (rather than the native driver that would see the 4 individual disks.) I couldn't get access to the data. Though it had worked when I was previously playing around with Xen (had tried copying it a 2 1TB RAID0 set, but then one of the 1TB drives I lost the copy, I had then replaced it with a 2TB RAID1 set....using an ST2000DL003 and an ST2000DM003, the DL being a 5900RPM drive and having a 5 year warranty...while the DM drive is a 7200RPM drive, but with only a 1 year warranty. And, turns out the 1 year is generous.

At work, I had built my FreeBSD desktop using a pair of the ST1000DM003 drives...and 3 drive failures is now a pair of ST2000DL003 drives. Yeah...I was having trouble with the array, and apparently using XFS was a mistake too...because I thought it was recovering, but instead it was slowly eating the data. When I had nuked the RR622 RAID5 array, and had switched to using it as JBOD and create a RAIDZ set under FreeBSD...I found that there was nothing to copy back from the RAID1 array. D'Oh! >:XX

Though I had also copied the Microsoft WindowsImageBackup files, to see if I could mount the VHD file under VirtualBox to help in recovery. I largely had the data in bits and pieces elsewhere, it was the environment I was wanting to recreate...and Oops!Backup didn't back up that part anyways (the data I was mainly trying to migrate). The image mounted, and I could see it...but soon after Windows would try to fix it and then it would disappear....kind of like what it did on February 15th to make the original Zen go away. No idea what kind of disk rotting the Intel Matrix RAID had been doing, when it had to initialize the array again every time after a Windows crash. I've had Ubuntu crashes, but the RAID arrays remained stable...usually. While Windows & Intel was pretty much every time. I'm sure it was slowly corrupting things overtime to where things wouldn't recover, though it choose to do that after an automatic reboot for Windows updates...and the day before I left for my first Gallifrey One made things even more annoying.

Anyway with another 1.5TB drive freed up, I contemplated adding it to the RAIDZ I had made of the 4 1.5TB drives, keeping it as a hot spare, or just use it by itself -- living dangerously. I ended up with the latter for some temporary data. Because in my mind I was starting to lean to what happened next.

Pages: 1· 2


  06:45:00 pm, by The Dreamer   , 478 words  
Categories: Hardware, Computer, Storage, Ubuntu

ulkc has 8GB of memory

So, a while back I read that it would be possible to upgrade a number of older computers from spec'd maximum to a bit more.

Officially, the maximum supported memory of the Lenovo 3000 V200 (ulkc) is 4GB... My other laptop also supports a maximum of 4GB, but I suspect that between them, I should be able to upgrade one or both to 8GB. Though I prefer to have it work in this one, since I use this one more.

So, a while back I had a newegg coupon to get laptop memory, so I got a pair of the SO-DIMMs. But, then I set them aside....and forgot about them.

More recently, I picked up a new battery for this laptop...and had it in the bedroom to run down and see how much life I would get. Online they sell aftermarket batteries for 4800mAH and 5200mAH. Though it turns out the original battery was 5200mAH.... So, that's what I have....anyways it works. And, I let the laptop rundown. So, I was going to try the memory swap after that. Though that was a few days ago....

Full story »


  10:06:00 pm, by The Dreamer   , 1662 words  
Categories: General, Hardware, Computer, Networking, BOINC, Storage, Ubuntu, Windows, Printers

lhaven misstep - nForce4 blows OCZ

So, lhaven was this computer I bought specifically for Stanford's Quake Catcher Network project. This was because when I received my sensor, I found that it only worked on Windows or Mac. Linux would require the older sensor, which they weren't selling anymore.

Things seemed to work...except that the machine would lock up now and then. Well, it is maybe I'll just schedule it to reboot regularly.

At first that seemed to help...but gradually, I was rebooting it more and more frequently. Eventually, I was down to rebooting it 5 times a week. Sometimes it would fail to restart, other times it would still lock up between reboots.

Eventually, they came out with Linux support for the sensor and I moved the sensor over to orac. That worked great for a while...and then it got a kernel update, and then starting VirtualBox would cause the machine to reboot. And, not just when I needed up update my Quicken. Because I was also participating in LHC's Test4Theory.

So, it was time to find a new home for the sensor. After moving it back to lhaven.

I looked for various low power x86 dedicated SBC or plug computer to run it. Though these things aren't cheap, so I'd want to do more than just QCN with it.

Then it occurred me...that I was overthinking this whole mess. Why don't I just install Linux on lhaven?

I haven't been looking forward to the leap from 10.04LTS to 12.04LTS. While Orac has a GT440 in it that's not being used, not sure what I would do for box....though I've been eye'ng either a GT218 or an NVS 300. It only has an LP PCIe x1 slot.... there are other concerns.

In the Scheme of things...I had planned to upgrade the video card in zen someday...and take its GT420 and change to an LP bracket and put it in lhaven. Hadn't quite worked out what I'll do about gumby, but it has since died...and when I get around to resurrecting it, I might have a Radeon HD5450 that I'll free up (not sure if it helps with BOINC, but eventually...I want all the systems using I can switch to my new 4-port DVI KVM. DVI to VGA through the old KVM was keyboard issues with the old KVM, makes me wonder if I can wait until I can fully make the switch.)

So, I had downloaded 12.04LTS beta1, but never got around to install it anywhere....and when this idea occurred, 12.04LTS beta2 was just starting to appear. Though the question was, would I go with ubuntu, or try one of the other variants. Namely, xubuntu or lubuntu.

I ended up going with xubuntu 12.04LTS beta 2. I didn't hear about lubuntu until after I had started downloading I envisioned that I would resurrect gumby with lubuntu....though so far I haven't gotten around to that.

Pages: 1· 2· 3


  03:27:00 pm, by The Dreamer   , 1112 words  
Categories: Software, Computer, Networking, Ubuntu, Windows, Printers, Scanner

I got a Brother DCP-7065DN

In a rather impulsive move, I bought a Brother DCP7065DN Monochrome Laser Multi-Function Copier via an Amazon Lightning Deal!

I had been considering getting a new laser printer for a very long time. I had ditched my old Brother HL-630 laser printer about a dozen years ago. This was a crappy printer from the start...with its grey background problem that they never seemed to resolve. Had gotten it from Future Shop of years way gone by....where it had been a special buy for they were selling them really cheap. Kind of like the Canon BJ-5e bubble jet printer I had gotten before.... (It was a BJ-10e, but made cheaper just for them. And, used a special ink cartridge not used by any other model from the time. Otherwise, I often wish I still had it....though when I moved from Canada to the US, it became kind of a problem to get ink for.)

So, after the Brother HL-630 was scrapped....I spent a long time thinking about my next laser printer....I was heavily leaning towards getting an HP Laser it was going to likely take a very long time before I would make a move. And, then I started to lean towards getting a Color Laser Printer.....and then I wanted built-in Duplex. And, having been spoiled by the HP Photosmart 8450xi, it also needed network capability....

There was an HP Color Laser printer that seemed to fit the bill, and it was popping up on sale now and then. But, then I upgraded to Windows 7...and found that not only is my old Photosmart 8450xi no longer fully supported in Windows 7, neither is the HP Color Laser printer that I had been considering.

So, that nixed wanting to jump on it the next time it went on sale. (Sale as in cheaper than it would cost to buy 4 new toner cartridges for it.)

So, the search continued....I started looking at Brother current models, something like the HL-4150CDN or the HL-4570CDWT.... but, then I was introduced to a multifunction printer. Which actually was kind of useful. Being able to scan a document onto a USB drive. And, there's always the once a year or so occasion, where somebody wants me to fax something. I used to have a fax machine...first a dedicated thermal fax type thing, and later a refurbished HP OfficeJet v40xi. A power surge took out my first fax machine...and then a few years in the same apartment still....a power surge took out the HP OfficeJet....

So, I then started to consider printers like the Brother the idea of color copying seemed useful, especially duplex color copying.

Of course, now its getting expensive to start and expensive to keep....

I also had this issue of printing envelopes....a capability I had lost when I upgraded to Windows 7. The Dymo Labelwriter that I had gotten from didn't work on Windows 7. And, I didn't really want to buy a new Dymo Labelwriter to keep this capability. Eventually I canceled, though still have the occasional postage need....which it would be nice if I could still print postage, etc. Think there are cheaper ways than, but probably not from Linux.

But, then last December while visiting my dad snapped up a Brother HL-2270DW, and this was a pretty decent little printer. It was on sale for the holidays...but it does PCL emulation. I wish I had just gotten one of these for myself then, too.

Instead, I started to reconsider whether I really needed color laser. Around year end...there were things I needed scanned or copied, for things like taxes and I was still was still leaning towards getting a multifunction printer of some sort. And, I kind of fell for the Samsung SCX-47xxFN, and then later the Samsung SCX-4828FN (about $400). The 4828FN adding Postscript emulation. I've always wanted a postscript printer....

Though these are still expensive printers. These ones had fax....and I did have to need to send a fax once this year, so far. If I were to leave out fax, it should've been something like the Brother DCP-8085DN (about $350)....which would've had duplex copying, and both PCL and PS emulations.

But, instead the Brother DCP-7065DN showed up as an Amazon Lightning deal for $80 (50% off)....I did a quick google search to see if it work on Linux, and some sites seemed to say it had PCL and PS as well. And, bought it.

Well...apparently its common in the industry to make similar internals/firmware across their product line with different features available or not, etc. And, Brother didn't even bother to remove PCL and PS from the capabilities page that the printer spits out or not advertise the LPD queues for these modes in its network info. But, trying to use the queues is like sending to /dev/null. This particular printer is only GDI (so in someways worse than the HL-2270DW ... currently about $120. And, then there's an HL-2280DW for about $10 more, and adds a scanner....a network scanner would be useful. Though a standalone scanner would be more useful during my situation where my primary desktop is dead.

But, I was able to get the DCP-7065DN working from my Ubuntu server, after following some websites on the subject. Namely..the detail that they only provided 32-bit deb's, and I have a 64-bit Ubuntu Server. I haven't tried the scanning feature, since the printer sits by TARDIS...and I have my Fujitsu ScanSnap S300M working on my Ubuntu server....though not sure if I've gotten the two-sided scanning part to work. So far the bills that I've needed to scan for FSA reimbursement have been one-sided.

But, the DCP-7065DN doesn't have a duplex it doesn't provide an alternative. And, my other scanner is a Fujitsu ScanSnap S300! (I bought the Mac only S300M on a Black Friday sale, and back when I used to bring work macbook pro home now and then. Later I got the S300, when I had more FSA money to spend and wanting faster turn around...but wasn't bring the macbook pro home anymore.) A regular flatbed scanner has been on my wishlist...perhaps the DCP-7065DN will satisfy that need. Something to probably check out after I get around to finally upgrading TARDIS.

So, while it was a problem to find that the DCP-7065DN is only a GDI does currently work from Ubuntu...and sharing it out from my Ubuntu server is interesting. There's no native DCP-7065DN driver in Windows XP, but I just pick some Postscript printer driver...and that worked. And, it looks pretty good (compared to the early days of ghostscript...)

What what I'll do when I finally get around to rebuilding my Windows 7 desktop????


  07:32:00 pm, by The Dreamer   , 693 words  
Categories: Hardware, Computer, Networking, Storage, Ubuntu

Time machine & NFS


So, during the long weekend...I got to wonder why I couldn't have my work MacBookPro do Time Machine backups to a network share, I had envisioned coming up with a drive to attach to one of my Airport Extremes and doing that some day. Though it doesn't look like that's actually going to happen anytime soon.

Did a quick google search on Time Machine and SMB, and found out about:

defaults write TMShowUnsupportedNetworkVolumes 1

And, so trying to make it work started. I created a Samba shared named 'SIDRAT' on ORAC...which I think is the name of the removable 160G drive that I had been using at home for Time Machine backups. The one at work is named 'TARDIS'. But, I already have a Samba shared named 'TARDIS' on ORAC....that's the 1TB volume for the HSTi Media Stick. Also not to be confused with computers named TARDIS and SIDRAT. Though SIDRAT died a long time ago, and I still haven't gotten around to replacing was my Windows laptop.

At first I didn't have dedicated storage on ORAC for this...I was just exporting the subdirectory of another Samba share as SIDRAT. And, it didn't really make sense for it to be a subdirectory on that share...but it was where I had at least 80GB to spare to this....the drive in the MacBookPro is only 80GB.

It seemed to start, but then it would fail. I had picked just the usual options for creating the Samba share, and wondered if the Mac needed special stuff to work. So, I went looking....

Read about the:


trick...but that didn't work. And, wasn't necessary when I did get it working. :))

So, then I read about creating the sparsebundle to get it started. There was conflicting info on how to name it....the usual was start Time Machine, and see what name it creates and use that.

So, was creating "<host>.tmp.sparsebundle" And, there were some variations on how to invoke hdiutil to create this. Well, it went slowly made it get smaller and smaller until it was gone and then went on to try to create "<host>.sparsebundle". Creating that as well didn't help. I had heard about "<host>_<MAC_ADDRESS>.sparsebundle", but the hits I found didn't explain just what MAC_ADDRESS would look like, or which one to use. Decided this wasn't going anywhere and the long weekend was nearly over....I did at least (after many long attempts) get a backuppc backup of parts of my MacBookPro. So, I got rid of the Samba share and undid most of the other stuff.

Full story »


  08:38:00 pm, by The Dreamer   , 601 words  
Categories: Hardware, Software, Computer, BOINC, Storage, Ubuntu

Got another brain for Orac

So, the other day, I was poking around on, trying to decide if I wanted to get the upcoming ShellShocker. I spotted a single slot GT440 for a good price among the daily deals. While I was looking around, I saw other things that caught my interest...a powersupply tester. Or things that I needed to go with the card, like a DVI-I to VGA adapter. Also decided to grab a new fan, since it is probably only a matter of time that the fan in the other drive array on Orac to fail.

I ended up not getting the ShellShocker item.

I've been eyeballing getting a single slot 400 Series card for Orac for a while. Was originally going to get a 200 Series, and I did actually get one in a different ShellShocker deal. But, when I opened up Orac, I realized it wouldn't fit....due to its huge fan.

Not sure what I'll do with that card. But, I had been looking at 400 Series, because Zen came with a GT420....which I've learned is a pretty horrible, compared to the rest of the line. Especially when it comes to doing BOINC. I'll probably go GTS450 or GTX460....single slot, though I need to take a look inside someday to see if there's actually juice in there. Otherwise, maybe I'll want a second one of these...

So, I had picked up the card from the UPS Store on Friday...but I didn't get around to installing it today. I stopped checkarray before I shutdown. I changed the scheduling of checkarray to spread things out better.

Though I had a brief distraction this morning on another matter, which I'll be posting about soon...

But, I went and put it in...its a pretty tight fit sitting next to my PM eSATA card.

I wonder if I'll look into a slot fan to go in the open PCI slot there....the lower PCI slot is where 4 other eSATA....oh wait, I had gotten a PCI card that I was going to put into Orac. I wonder what happened to that project....

Pages: 1· 2


  09:48:00 am, by The Dreamer   , 892 words  
Categories: Hardware, Computer, BOINC, Roku XDS, Ubuntu, Other Linux

LHAVEN is dead, long live LHAVEN?

Well, I was getting really annoyed with it complaining and kicking out the replaced disk in LHAVEN. So, I shutdown the system and took out the disk, figuring it should just continue to run fine degraded until I get around to replacing it.

Well, it wouldn't power on after that. I went through everything, no joy. It had done this a while back when I tried adding something to it. It could just be the power supply has flaked out, but I don't have a tester or multimeter I can't really test it. Could buy a new PS...and I may do that.

But, I kind of suspected the drive issues weren't actually the drive but possibly some deeper hardware problem. So, I had been planning to replace LHAVEN at some point. It has done quite well, being circa 2002. Started out as a 64MB Duron 800MHz machine, eventually peaked at 2GB Athlon XP 3000+. It had replaced a Cyrix PR233 box that had gone up in smoke during a hot summer day in 2002, when the transformer outside blew....the computer survived the brownout and then blackout, but the fans didn't spin back up when power it burned itself up. Later I found one problem with this new system. It wouldn't resume after losing power. No BIOS setting to alter this behavior and I did try to see if there was alternate BIOS updates for it. It was kind of a painful machine to manage, because for some time kernels didn't have built-in support for all the SIS chipset stuff in an update would roll out, and NIC was usually one of the things to definitely go missing. Which made for fun to rebuild a custom kernel upgrade for it. Eventually it got stable.

But, needing essentials like DNS, DHCP to be available after an extended outage...I moved these to another server (originally an old Pentium 75, which has evolved into what is known as 'box' today). For the longest time it was RedHat 7.2, and then RedHat 7.3 when fedora legacy switched to only supporting 7.3 and 9 releases. It continued after fedoralegacy stopped supporting it, I was building some of the packages for it by bind (in response to the Kaminsky exploit). There were parts of my network that wouldn't function without an old Windows 2000 box that was barely functioning...(old Gumby).

I nearly lost it during the Icepocalyse....but I tracked down motherboard replacement for it from It was a slighty newer mobo, but equivalent chipset. It did have some things the old mobo didn't, like USB 2.0 support (I didn't use USB and still don't), support for 2GB of RAM instead of I upped it for better BOINC'ng. And, faster I upped from 2200 to 3000.

At one time, I had turned it off to do some upgrades...which didn't pan out. Tried to slap a gigabit card into it...didn't work. Wouldn't get along with the on board stuff, and the BIOS didn't have ways to get things out of the way enough. It also never fixed the resume after power loss issue. But, during this process it failed to power on....but after a couple days, it came back and I decided to upgrade its UPS and hopefully it would make one last outage. Well, there were several others since then where it came back afterwards. But, looks like this is the last time for it.

Pages: 1· 2

  12:03:00 am, by The Dreamer   , 1362 words  
Categories: Software, Computer, Networking, WiFi, Ubuntu

Freeradius & DHCP Failover


So, ever since I looked at adding Mac Address Access Controls to my Airport Extreme...on top of WPA2 Personal, and the fact that my DHCP server only does reserved IPs, security. I used to do Mac Address Access Controls on my previous routers, but it was an easier interface to work with on those. And, I didn't realize how the Time Access worked on the Airport Extreme, the default allow all the time rule at the top tripped me up. So, I thought if I wanted it, I would need a RADIUS server...and I didn't know if I wanted to do that....yet.

But, after I woke one morning and couldn't seem to account for why there seemed to be so much data streaming through my Cox connection...there had been strange spikes in the past, but always figured it was something updating itself while I wasn't home (like iTunes and my podcast subscriptions). But, this one morning...there was no corresponding activity from any of my computers, and I didn't see anything obvious with my TiVos/ReplayTVs. Though I could've just missed it.

So, I fixed the Timed Access control and put my current devices in. With a note that I should really look into installing RADIUS somewhere, so that it would be easier to maintain the list than the airport utility. I would lose being able to find the MAC address of some new wireless device that doesn't have the MAC address stamped on it....for addition to my DHCP server.

Later during the setup in: Another Airport comes to Lunatic Haven I had wiped out the settings....and didn't feel like putting it back in again. Which made it more urgent (in my mind) to get RADIUS working.

So, I went online and searched and searched and searched...on how to do this. I had looked before, and wasn't all that successful. There's no simple how-to apparently. But, I found bits and pieces around, and decided to just go for it.

First, I installed freeradius on my Ubuntu server 'box'.

sudo apt-get install freeradius

It starts right away, now to make it work. And, debug it. Well, most of the examples were for older freeRADIUS versions, so things weren't where it said, or command line switches were different, or it didn't work. I did find some examples of MAC address authorization, but they involved 'Auth-Type := Local' in the /etc/freeradius/users file. But, the clients.conf part seemed right. I strongly considered just doing 'Auth-Type := Accept'...but I wanted to figure this mess out.

client {
        secret = testing123
        shortname = airport
        nastype = other

So kept searching and searching....eventually, I found fragments on site called "Deploying RADIUS: Practices and Principles". It confirmed that I was basically on the right track, I just needed to figure out what to put in the users file to make it go from Auth-Reject to Auth-Accept.

Well, the example for MAC Address entry for users I had found was:

001122-334455  Auth-Type := Local,  User-Password == "testing123"

At first I was pointing my Airport Extreme at it and watching the debug output, and watching everything stop working now and then. But, eventually I used 'radtest' to test my freeRADIUS configuration. And, eventually, I found that what I needed was:

001122-334455  Cleartext-Password := "testing123"

And, all was good. I pointed my main Airport Extreme to it, and it everything adjusted and worked. I then pointed the new Airport Extreme at it and things continued to work.

Yay! &#58;&#99;&#111;&#111;&#108;&#58;

Pages: 1· 2


  07:57:00 am, by The Dreamer   , 1339 words  
Categories: Hardware, Computer, Storage, Ubuntu

Worked on Orac last night

Back on March 28th, when I last wrote about working on Orac, I mentioned that I looked at the harddrive cage to see about the condition of the fan on there....only to find that there was no fan there.

Reviewing the manual on the Gateway website, I found that the cage is used in more than one model...and that some of those models have fans, while evidently mine did not. Browsing the parts list for some of the other Gateway models that used the same drive cage, I found reference to a 60mm x 10mm fan, which I deduced was probably the fan that I would need to get for this location. I did find that from the Gateway manual for my model, that the motherboard did have a front chassis fan connector.

So, after some thought and checking first, it struck me that eBay might be the better place to go. So I found a seller on eBay that explicitly said he shipped by USPS and bought one, and from another seller I got fan screws (a bunch of them, because I've needed them in the past before and I'm sure I'll have need for them in the it was I didn't actually need them this time though.)

Because I had recently built my new backuppc pool (should be posting about that adventure some day), I had been waiting for a moment when Orac was idle again and not busy refilling the pool with full backups of everything....It hadn't gotten any fulls of Zen yet, it didn't detect that Zen had gone away to apply the recent Microsoft patches during its recent attempt, so I had to step in and stop it. So that seemed like a good time to take Orac down.

Pages: 1· 2· 3· 4


  01:40:00 pm, by The Dreamer   , 389 words  
Categories: Software, Networking, Cox HSI, AT&T DSL, Ubuntu

ddclient & squid

In the aftermath of the summer storm of August 13th, (hmmm, totally missed that it was a Friday the 13th), I made a tweak to my ddclient config for updating dyndns for my DSL line. Because I found that it wasn't able to update the IP change while Cox was down.

Couldn't find a way to make ddclient to bind to the local IP that routes out by DSL (or use non-default gateway). But, since I have squid proxy on the same box...and depending on what port I come in on, it can use either of my connections.

I set

Couldn't use localhost, because ddclient does some kind of validation to require an fqdn+port, and localhost isn't an fqdn. And, yes, I use my dyndns domain as my home domain. So I can have bookmarks that'll work whether I'm at home or on the road &#59;&#68;

But, this change wasn't it has been less than 28 days for a refresh, and no IP change.

That was until this morning, when my IP did change.

The updates weren't working....seems that ddclient wants to do SSL all the way or not at all. No using an http proxy to connect out on SSL. But, I didn't feel like sending my dyndns password out non-SSL.... So, after some thought, I decided I would figure out how to set up SSL on squid.

I made the necessary configuration change, but no go. Seems that ubuntu doesn't distribute squid with SSL, because squid and openssl have incompatible open source licenses. So, I did a quick search to find the ubuntu way of rebuilding it from source.

apt-get source squid
apt-get build-dep squid
apt-get install devscripts build-essential fakeroot
cd squid-2.7.STABLE7
vi debian/rules
     Add --enable-ssl \ to “# Configure the package” section
debuild -us -uc -b
cd ..
dpkg -i squid??? squid-common???

Change to, and it worked. &#58;&#99;&#111;&#111;&#108;&#58;

Full story »

2 4 5

Now instead of subjecting some poor random forum to a long rambling thought, I will try to consolidate those things into this blog where they can be more easily ignored profess to be collected thoughts from my mind.

Latest Poopli Updaters --


There are 19 years 3 months 23 days 13 hours 25 minutes and 7 seconds until the end of time.
And, it has been 5 years 9 months 4 days 37 minutes and 49 seconds since The Doctor saved us all from the end of the World!


September 2018
Mon Tue Wed Thu Fri Sat Sun
 << <   > >>
          1 2
3 4 5 6 7 8 9
10 11 12 13 14 15 16
17 18 19 20 21 22 23
24 25 26 27 28 29 30


  XML Feeds

Who's Online?

  • Guest Users: 2
This seal is issued to by StopTheHacker Inc.
powered by b2evolution

hosted by
Green Web Hosting! This site hosted by DreamHost.

monitored by
Monitored by eXternalTest
SiteUptime Web Site Monitoring Service
website uptime