Dump, tar and friends are reliable and well understood but could be more flexible and efficient. In a modern server infrastructure you may be backing up hundreds of physical and virtual machines. Much of the data is likely to be repeated many times. A deduplicating archiver can be used to store only the unique information.
This talk covers our experiences with the Borg Deduplicating Archiver.
The talk we will discuss the topic of when it is best to start monitoring your products. Open source monitoring and cloud architecture make starting your monitoring that much easier and cheaper then most people think or are aware of. We will talk about why and when you may want to start monitoring, talk about the common mistakes present in many places in the industry, and we will discuss why it needs to change, and how to change it. And we will show case the ways that Icinga can help you adopt early monitoring and implementation in your work place.
Jenkins is a tool more and more businesses rely on. To build, test, and deploy code and pieces of infrastructures. But how do you automate and scale that piece of infrastructure? Automating Jenkins has been a pain in the past, but it is becoming better nowadays, with the introduction of Jenkins Pipeline and exposure of Jenkins Internal API’s.
This talk is a dive into the Jenkins world and its plugins.
The art and science of reconstructing a chain of events on digital systems – has for a long time been dominated by commercial software packages – EnCase, FTK, X-Ways and so on. With a small number of exceptions, when you are taught the subject, you’ll start on a Windows workstation. The argument historically had been that as you are likely to be examining a Windows device ( market prevalence of Windows in office and home environments ) familiarity with and the compatibility of the Windows based tool chain was the most logical way forward. Well, times have changed. It may or may not be the year of the Linux Desktop ( again ! ) but ChromeOS, Android, MacOS X ( now macOS ) and BSD based devices ( PS4 ) are turning up more and more often in real world cases. There are more strong tools available, more libraries to support tool development and a great justification in court for the transparency of methods used to obtain evidence. Is now the time to make the switch to Open Source?
Every once in a while, the proverbial hits the fan. There are so many ways that this can be the case that are nothing to do with the fault of the poor Systems Administrator – be it end user related or malicious 3rd party. In some really unfortunate scenarios, it may be that you’ll have to interface with Law Enforcement or external forensic analysts – this talk is about ensuring the preservation of evidence and working with those 3rd parties to enable your systems to keep working and to bring the “bad guys” to justice…
The origin behind Icinga and the evolution of the product, from both the code stand point and the user experience and input, changes adaptation and adoption in the world of monitoring.
The Nagios Plugin API has become a well loved standard for providing scripts that act as plugins to monitoring systems, and along with the standard plugins supplied by http://monitoring-plugins.org/ (formerly “Nagios Plugins” prior to legal threats from Nagios Inc) many more are available on various exchanges and on github.
Unfortunately, a noticeable portion of the runtime of some scripts is due to library load overhead because of the single shot runtime model, and many more such scripts eschew library use in favour of copypasta to avoid overhead and dependency management issues.
Both the deployment and scalability issues are solvable, however, at least for perl and CPAN.
Firstly, this talk will cover how to keep deployment simple while still being able to use CPAN modules, including:
* How to track your module use to ensure you always know not only what your dependencies are, but can reproduce the exact versions you’ve developed and tested against
* How to turn a plugin with pure perl dependencies – including the monitoring-plugins.org standard library Monitoring::Plugin – into a single
file that can be deployed without additional installation effort
* How to handle compiled dependencies both via easy in-tree installation to avoid touching system directories and via building system packages for global installation
Secondly, we’ll discuss the limitations of the execution model, and the issues and trade-offs involved in various approaches to a more persistent execution model. Having covered the previous solutions, and why they often don’t end up getting used, we’ll present an alternative that:
* Requires zero changes to plugin code
* Intelligently preloads modules to provide fast warming
* Shares dependency loading to minimise memory usage
* Can detect and work around broken modules automatically
* Is robust to even the most legacy or newbie perl scripting
Come along and find out how to both save yourself time by accelerating your development process, and save your systems’ CPU by accelerating your deployments. Perl experience is not expected, because it *is* expected that these techniques should apply to third party/community sourced code as well.
An overview of illumos six years later.
Just prior to Oracle closing OpenSolaris, the illumos project forked from it, and for six years has carried on the work started with OpenSolaris.
In August 2010, Garrett D’Amore and others forked OpenSolaris with the intention of reducing and eliminating its closed-source components. Unknown to them, Oracle was planning to shut off publication of the OS/Net consolidation, which included the OpenSolaris kernel & system-critical user-space. This coincidence propelled the fork, illumos, into the vehicle in which OpenSolaris could continue, albeit under a different name.
illumos is now over six years old, and its distributions continue to fulfill the promise of OpenSolaris. illumos is the reference implementation of OpenZFS and DTrace. SmartOS has continued pioneering work in zones, commonly known outside Solarish circles as containers. Both OpenIndiana and OmniOS use the Image Packaging System (IPS) in traditional deployments. OpenIndiana and XStreamOS continue work on an illumos Desktop. DilOS (Debian) and Tribblix (SVr4) use alternative packaging with illumos. Appliance distributions like
NexentaStor and Delphix use and innovate with OpenZFS to great effect.
The oVirt Project is an open virtualization project providing a feature-rich server and desktop virtualization management platform with advanced capabilities for hosts and guests, including high availability, live migration, storage management, system scheduler, and more.
oVirt provides an integration point for several open source virtualization technologies, including kvm, libvirt, spice and oVirt node.
oVirt was launched in November 2011 as a fully open source project, based on assets from Red Hat Virtualization Manager platform. The project has an open governance model, and initial board has members from IBM, Cisco, Netapp, Red Hat and SUSE.
The session will provide an intro to the project components and features.
KDE is the original and best community making end-user software on Linux and other platforms. However the techniques used to deploy much of this software have not updated well with a large lag between releases and the ability of users to get hold of the software.
KDE neon is brining a revolution to the way open source delivers its software on Linux. A KDE project providing KDE software directly to our users cuts the feedback loop and allows us complete control of the experience.
Using the latest buzzword compliant devops techniques with containerised continuous integration and continuous deployment, cloud provisioning and orchestration with an active engagement through social media presence, KDE neon is updating the model of how KDE provides its software.
We will show how KDE neon is used by Plasma developers to test their latest software and their releases. We’ll show how people interested in trying KDE’s software finally have an answer to the questions of where to get up to date versions.
And you can finally find an answer to the question, KDE neon: terrible idea, or simply a huge mistake?
Our department was located in five then four buildings and moved to a new home at the start of 2016. This talk will cover the successful relocation of all servers and core networking to a new data centre with minimal downtime, followed by the migration of the department to its new home. With over a thousand Cat 6A network outlets, dark fibre to the data centre, new VoIP phone system, comprehensive Wi-Fi coverage, and hundreds of users with desktops, laptops and a few printers all to be installed. The successful migration of home directories will then be described moving away from an old Sun SPARC Solaris system to modern day Dell Intel servers using Debian Linux, with ext4fs, LVM2, NFS, Samba and ZFS.
Lets Encrypt has become a much talked about system, however its recommended software hooks into a lot of different parts, and requires a lot more access than I was happy with. Take a quick look at a set of scripts and framework I created for setting up and fetching SSL keys, and secure your public sites for free with ease!
Eighteen months ago I decided to give up the relative security of a permanent position at a large university and become a swashbuckling mercenary, selling my development and system administration skills to the highest bidder. Unfortunately the reality turned out to be slightly less glamorous and I’ve ended up as a coin-operated monkey, but I’ve had lots of fun along the way.
This talk will cover my experiences of freelancing so far, whether it might be the right choice for you, and advice on what (and what not) to do if you decide to join us on the dark side (we have cookies).
When you’re testing software, you want the framework to test both individual units, as well as the whole application. However, sometimes you require external processes, such as Apache, Postgres, or other such software. Setting up a test environment to include these can be difficult – or so the case used to be. Join me for a journey to see how such a module could be created, and how you can build frameworks in perl which will test anything.
Moving from the Iron Age to the Cloud Age in computing is supposed to save us money yet many migrations seem to cost more in the long run and result in infrastructures as complex to manage as what we had before. This is often the result of the so called “lift & shift” approach many take – it’s a short term win that doesn’t address why you wanted to move to the cloud in the first place.
The Cloud Age affords us the opportunity not to treat our infrastructure as something special, instead as something disposable. By applying the practices of continuous integration and delivery to our infrastructure and configuration management we can built truly scalable infrastructures to host our application’s wildest dreams.
In this talk we will look at the tools and processes that can be adopted to truly make use of the possibilities of the Cloud.