Language Selection

English French German Italian Portuguese Spanish

Site Map


Community blog and recent blog authors at Tux Machines.





More in Tux Machines

Videos/Audiocasts/Shows: GNU/Linux and Python, Fresh Look at LMDE 4 Beta

  • Hopeful for HAMR | TechSNAP 423

    We explore the potential of heat-assisted magnetic recording and get excited about a possibly persistent L2ARC. Plus Jim's journeys with Clear Linux, and why Ubuntu 18.04.4 is a maintenance release worth talking about.

  • 2020-02-21 | Linux Headlines

    Red Hat OpenStack Platform reaches version 16, Google announces the mentors for this year’s Summer of Code, DigitalOcean secures new funding, the Raspberry Pi 4’s USB-C power problems get a fix, and the GTK Project unveils its new website.

  • Talk Python to Me: #252 What scientific computing can learn from CS

    Did you come into Python from a computational science side of things? Were you just looking for something better than Excel or Matlab and got pulled in by all the Python has to offer?  That's great! But following that path often means some of the more formal practices from software development weren't part of the journey.  On this episode, you'll meet Martin Héroux, who does data science in the context of academic research. He's here to share his best practices and lessons for data scientists of all sorts.

  • Matt Layman: Templates and Logic - Building SaaS #45

    In this episode, we added content to a template and talked about the N+1 query bug. I also worked tricky logic involving date handling. The first change was to update a course page to include a new icon for any course task that should be graded. After adding this, we hit an N+1 query bug, which is a performance bug that happens when code queries a database in a loop. We talked about why this happens and how to fix it. After finishing that issue, we switched gears and worked on a tricky logic bug. I need a daily view to fetch data and factor in the relative time shift between the selected day and today. We wrote an involved test to simulate the right conditions and then fixed the code to handle the date shift properly.

  • LMDE 4 Beta Debbie Run Through

    In this video, we are looking at LMDE (Linux Mint Debian Edition) 4 Debbie.

KVM and Xen Project: Commercial Exploitation and Unikraft Work

  • Cloud, Linux vendors cash in on KVM-based virtualization

    Vendors such as Red Hat, IBM, Canonical and Google rely on KVM-based virtualization technology for many of their virtualization products because it enables IT administrators to execute multiple OSes on the same hardware. As a result, it has become a staple in IT admins' virtual systems. KVM was first announced in October 2006 and was added to the mainline Linux kernel in February 2007, which means that if admins are running a Linux machine, they can run KVM out of the box. KVM is a Type 1 hypervisor, which means that each individual VM acts similar to a regular Linux process and allocates resources accordingly. Other Type 1 hypervisors include Citrix XenServer, Microsoft Hyper-V, Oracle VM Server for x86 and VMware ESXi.

  • Unikraft: Building Powerful Unikernels Has Never Been Easier!

    Two years ago, the Xen Project introduced Unikraft ( as an incubation project. Over the past two years, the Unikraft project has seen some great momentum. Since the last release, the community has grown about 20% and contributions have diversified a great deal. Contributions from outside the project founders (NEC) now make up 63% of all contributions, up from about 25% this time last year! In addition, a total of 56,739 lines were added since the last release (0.3). [...] Finally, the Unikraft team’s Simon Kuenzer recently gave a talk at FOSDEM titled “Unikraft: A Unikernel Toolkit”. Simon, a senior systems researcher at NEC Labs and the lead maintainer of Unikraft, spoke all about Unikraft and provided a comprehensive overview of the project, where it’s been and what’s in store.

Gopher: When Adversarial Interoperability Burrowed Under the Gatekeepers' Fortresses

In the early 1990s, personal computers did not arrive in an "Internet-ready" state. Before students could connect their systems to UMN's network, they needed to install basic networking software that allowed their computers to communicate over TCP/IP, as well as dial-up software for protocols like PPP or SLIP. Some computers needed network cards or modems, and their associated drivers. That was just for starters. Once the students' systems were ready to connect to the Internet, they still needed the basic tools for accessing distant servers: FTP software, a Usenet reader, a terminal emulator, and an email client, all crammed onto a floppy disk (or two). The task of marshalling, distributing, and supporting these tools fell to the university's Microcomputer Center. For the university, the need to get students these basic tools was a blessing and a curse. It was labor-intensive work, sure, but it also meant that the Microcomputer Center could ensure that the students' newly Internet-ready computers were also configured to access the campus network and its resources, saving the Microcomputer Center thousands of hours talking students through the configuration process. It also meant that the Microcomputer Center could act like a mini App Store, starting students out on their online journeys with a curated collection of up-to-date, reliable tools. That's where Gopher comes in. While the campus mainframe administrators had plans to selectively connect their systems to the Internet through specialized software, the Microcomputer Center had different ideas. Years before the public had heard of the World Wide Web, the Gopher team sought to fill the same niche, by connecting disparate systems to the Internet and making them available to those with little-to-no technical expertise—with or without the cooperation of the systems they were connecting. Gopher used text-based menus to navigate "Gopherspace" (all the world's public Gopher servers). The Microcomputer Center team created Gopher clients that ran on Macs, DOS, and in Unix-based terminals. The original Gopher servers were a motley assortment of used Macintosh IIci systems running A/UX, Apple's flavor of Unix. The team also had access to several NeXT workstations. Read more Also: The Things Industries Launches Global Join Server for Secure LoRaWAN

IBM/Red Hat and POWER9/OpenBMC

  • Network Automation: Why organizations shouldn’t wait to get started

    For many enterprises, we don’t need to sing the praises of IT automation - they already get it. They understand the value of automation, have invested in a platform and strategy, and have seen first-hand the benefits IT automation can deliver. However, unlike IT automation, according to a new report from Forrester Research 1, network automation is still new territory for many organizations. The report, "Jump-Start Your Network Automation," found that 56% of global infrastructure technology decision makers have implemented/are implementing or are expanding/upgrading their implementation of automation software, while another 19% plan to implement it over the next 12 months. But those same organizations that are embracing IT automation haven’t necessarily been able to take that same initiative when it comes to automating their networks. Even if they know it will be beneficial to them, the report found that organizations often struggle with even the most basic questions around automating their networks.

  • Using a story’s theme to inform the filmmaking: Farming for the Future

    The future of farming belongs to us all. At least that’s the message I got from researching Red Hat’s most recent Open Source Stories documentary, Farming for the Future. As a self-proclaimed city boy, I was intrigued by my assignment as director of the short documentary, but also felt like the subject matter was worlds away. If it did, in fact, belong to all of us how would we convey this to a general audience? How could we use the film’s theme to inform how we might approach the filmmaking to enhance the storytelling?

  • Raptor Rolls Out New OpenBMC Firmware With Featureful Web GUI For System Management

    While web-based GUIs for system management on server platforms with BMCs is far from anything new, Raptor Computing Systems with their libre POWER9 systems does now have a full-functioning web-based solution for their OpenBMC-powered systems and still being fully open-source. As part of Raptor Computing Systems' POWER9 desktops and servers being fully open-source down to the firmware/microcode and board designs, Raptor has used OpenBMC for the baseboard management controllers but has lacked a full-featured web-based system management solution on the likes of the Talos II and Blackbird systems up until now.

  • Introduction to open data sets and the importance of metadata

    More data is becoming freely available through initiatives such as institutions and research publications requiring that data sets be freely available along with the publications that refer to them. For example, Nature magazine instituted a policy for authors to declare how the data behind their published research can be accessed by interested readers. To make it easier for tools to find out what’s in a data set, authors, researchers, and suppliers of data sets are being encouraged to add metadata to their data sets. There are various forms for metadata that data sets use. For example, the US Government site uses the standard DCAT-US Schema v1.1 whereas the Google Dataset Search tool relies mostly on tagging. However, many data sets have no metadata at all. That’s why you won’t find all open data sets through search, and you need to go to known portals and explore if portals exist in the region, city, or topic of your interest. If you are deeply curious about metadata, you can see the alignment between DCAT and in the DCAT specification dated February 2020. The data sets themselves come in various forms for download, such as CSV, JSON, GeoJSON, and .zip. Sometimes data sets can be accessed through APIs. Another way that data sets are becoming available is through government initiatives to make data available. In the US, has more than 250,000 data sets available for developers to use. A similar initiative in India,, has more than 350,000 resources available. Companies like IBM sometimes provide access to data, like weather data, or give tips on how to process freely available data. For example, an introduction to NOAA weather data for JFK Airport is used to train the open source Model Asset eXchange Weather Forecaster (you can see the model artifacts on GitHub). When developing a prototype or training a model during a hackathon, it’s great to have access to relevant data to make your solution more convincing. There are many public data sets available to get you started. I’ll go over some of the ways to find them and provide access considerations. Note that some of the data sets might require some pre-processing before they can be used, for example, to handle missing data, but for a hackathon, they are often good enough.

  • Red Hat Helps Omnitracs Redefine Logistics And Transportation Software

    Fleet management technology provider Omnitracs, LLC, has delivered its Omnitracs One platform on the foundation of Red Hat OpenShift. Using the enterprise Kubernetes platform along with Red Hat Ansible Automation Platform, Omnitracs One is a cloud-native offering and provides an enhanced user experience with a clear path towards future innovations. With Red Hat’s guidance, Omnitracs said it was able to embrace a shift from on-premises development technologies to cloud-native services, improving overall operations and creating a more collaborative development process culture.