Nov 172015

Article By Cassie Philips.

Linuxaria has been one of my go-to sources for Linux tips and information ever since I discovered it a couple of months ago. I would strongly suggest reading this post on managing processes with cgroup, as it is an example of the quality of their content.

Linux being one of the quickest rising operating systems available to computers (especially with the privacy debacle that is Windows 10), we are finding that more tools are becoming available to Linux users for their convenience, entertainment and security. A Virtual Private Network (VPN) is one of those services.

What is a VPN, and how does it relate to you? Those are the questions this article is here to answer for you.

Continue reading »

Flattr this!

Jul 102015

This is a re-post of the original article byJason SkowronskiAmy Echeverri and Sadequl Hussain first posted on

Linux Logging Basics

First we’ll describe the basics of what Linux logs are, where to find them, and how they get created. If you already know this stuff, feel free to skip to the next section.

Linux System Logs

Many valuable log files are automatically created for you by Linux. You can find them in your /var/log directory. Here is what this directory looks like on a typical Ubuntu system:


Some of the most important Linux system logs include:

  • /var/log/syslog or /var/log/messages stores all global system activity data, including startup messages. Debian-based systems like Ubuntu store this in /var/log/syslog. RedHat-based systems like RHEL or CentOS store this in /var/log/messages.
  • /var/log/auth.log or /var/log/secure stores logs from the Pluggable Authentication Module (pam) including successful logins, failed login attempts, and authentication methods. Ubuntu and Debian store authentication messages in /var/log/auth.log. RedHat and CentOS store this data in /var/log/secure.
  • /var/log/kern stores kernel error and warning data, which is particularly helpful for troubleshooting custom kernels.
  • /var/log/cron stores information about cron jobs. Use this data to verify that your cron jobs are running successfully.

Continue reading »

Flattr this!

May 292014

Today I want to repost for my readers a really interesting article by Gionatan Danti first posted on his blog, I hope you enjoy it as much as I do

File compression is an old trick: one of the first (if not the first) program capable of compressing files was “SQ”, in the early 1980s, but the first widespread, mass-know compressor probably was ZIP (released in 1989).

In other word, compressing a file to save space is nothing new and, while current TB-sized, low costs disks provide plenty of space, sometime compression is desirable because it not only reduces the space needed to store data, but it can even increase I/O performance due to the lower amount of bits to be written or read to/from the storage subsystem. This is especially true when comparing the ever-increasing CPU speed to the more-or-less stagnant mechanical disk performance (SSDs are another matter, of course).

While compression algorithms and programs varies, basically we can distinguish to main categories: generic lossless compressors and specialized, lossy compressors.

If the last categories include compressors with quite spectacular compression factor, they can typically be used only when you want to preserve the general information as a whole, and you are not interested in a true bit-wise precise representation of the original data. In other word, you can use a lossy compressor for storing an high-resolution photo or a song, but not for storing a compressed executable on your disk (executable need to be perfectly stored, bit per bit) or text log files (we don’t want to lose information on text files, right?).

So, for the general use case, lossless compressors are the way to go. But what compressor to use from the many available? Sometime different programs use the same underlying algorithm or even the same library implementation, so using one or another is a relatively low-important choice. However, when comparing compressors using different compression algorithms, the choice must be a weighted one: you want to privilege high compression ratio or speed? In other word, you need a fast and low-compression algorithm or a slow but more effective one?

In this article, we are going to examine many different compressors based on few different compressing libraries:

  • lz4, a new, high speed compression program and algorithm
  • lzop, based on the fast lzo library, implementing the LZO algorithm
  • gzip and pigz (multithreaded gzip), based on the zip library which implements the ZIP alg
  • bzip2 and pbzip2 (multithreaded bzip2), based on the libbzip2 library implementing the Burrows–Wheeler compressing scheme
  • 7-zip, based mainly (but not only) on the LZMA algorithm
  • xz, another LZMA-based program

Continue reading »

Flattr this!

arkOS: build your Cloud with a  Raspberry Pi

arkOS: build your Cloud with a Raspberry Pi

Recently I’ve discovered this project that has great ambitions: arkOS is an open-source platform for securely self-hosting your online life. Everything started from the founder Jacob Cook and the CitizenWeb Project he founded. It’s designed to run on a Raspberry Pi – a super-low-cost single board computer – and ultimately will let users, even of […]

Hadoop – The Small Application for Big Data

Back in December 2011, data-intensive Linux users rejoiced as Apache Hadoop reached its 1.0.0 milestone. Setting a benchmark for distributed computing software, this wonderful little program is now into release 1.0.3 but what is Hadoop and how can you benefit from using it? Designed with ‘web-scale’ operations in mind, Hadoop can handle massive amounts of […]