This document introduces Project Builder, a tool for continuously packaging software projects. It allows packaging to be done in an operating system agnostic way across different Linux distributions and package formats. Project Builder uses container technologies like Docker to provide isolated build environments for projects. It aims to make packaging easier for upstream projects and developers to integrate with various distribution repositories and deployment systems.
Project-Builder.org is a tool to help upstream projects build their own packages for various Operating Systems directly from their VCS/CMS on a regular basis. It provides a continuous packaging approch
Building distribution packages with DockerBruno Cornec
This presentation demonstrate how to use Docker in order to build upstream Fedora or Mageia distribution packages much more easily than usual. It was given during Fosdem 2016
Presentation on HP ProLiant value add tools on LinuxBruno Cornec
The document discusses HP's support for Linux and open source software. It provides an overview of HP's strategy and commitments to open source, including consuming open source in products and IT, contributing to communities, and supporting communities. It also describes HP's support for Linux on ProLiant servers, including supporting various distributions, the Software Delivery Repository for drivers and tools, and the Management Component Pack.
Lab Document on HP ProLiant value add tools on LinuxBruno Cornec
This document provides an overview of various command line tools for managing HP ProLiant servers running Linux, including hplog, hpuid, hpasmcli, hpbootcfg, and hponcfg. It describes how to use these tools to view hardware logs, control the server identify light, modify boot settings, enable/disable automated server recovery, and configure settings on the iLO management processor like the system name. The tools are packaged as hp-health, hpssacli, and hponcfg for installation on supported Linux distributions.
This document discusses HP's involvement with open source software and Linux. It notes that HP has supported Linux on its servers for over 14 years through driver and firmware updates. It highlights some of HP's strategic partnerships with companies like Intel, Red Hat, and OpenStack. The document also summarizes tools and utilities HP provides to manage its servers running Linux, such as the Management Component Pack and Service Pack for ProLiant.
The document discusses HP's support for Linux on its ProLiant servers. It describes how HP tests commercial Linux distributions on ProLiant hardware, enables certification of distributions near new hardware launches, and provides the ProLiant Support Pack. It also discusses HP's Software Delivery Repository, which provides drivers, utilities, and tools. Finally, it explains HP tools like iLO, the Management Component Pack, and Service Pack for ProLiant, which provide additional monitoring, management, and update capabilities for ProLiant Linux systems.
In this deck from the 2016 HPC Advisory Council Switzerland Conference, Kenneth Hoste from the University Ghent presents an introduction to EasyBuild, an open-source framework for (automatically) getting scientific software installed on HPC systems.
Watch the video presentation: http://wp.me/p3RLHQ-f8J
Learn more: https://github.com/hpcugent/easybuild
See more talks from the Switzerland HPC Conference:
http://insidehpc.com/2016-swiss-hpc-conference/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Yocto Project Dev Day Prague 2017 - Advanced class - Kernel modules with eSDKMarco Cavallini
Excerpt of the lectures at Yocto Project Dev Day in Prague, 2017
During the advanced class Marco Cavallini ran this presentation about the creation of a Linux kernel module with eSDK.
Project-Builder.org is a tool to help upstream projects build their own packages for various Operating Systems directly from their VCS/CMS on a regular basis. It provides a continuous packaging approch
Building distribution packages with DockerBruno Cornec
This presentation demonstrate how to use Docker in order to build upstream Fedora or Mageia distribution packages much more easily than usual. It was given during Fosdem 2016
Presentation on HP ProLiant value add tools on LinuxBruno Cornec
The document discusses HP's support for Linux and open source software. It provides an overview of HP's strategy and commitments to open source, including consuming open source in products and IT, contributing to communities, and supporting communities. It also describes HP's support for Linux on ProLiant servers, including supporting various distributions, the Software Delivery Repository for drivers and tools, and the Management Component Pack.
Lab Document on HP ProLiant value add tools on LinuxBruno Cornec
This document provides an overview of various command line tools for managing HP ProLiant servers running Linux, including hplog, hpuid, hpasmcli, hpbootcfg, and hponcfg. It describes how to use these tools to view hardware logs, control the server identify light, modify boot settings, enable/disable automated server recovery, and configure settings on the iLO management processor like the system name. The tools are packaged as hp-health, hpssacli, and hponcfg for installation on supported Linux distributions.
This document discusses HP's involvement with open source software and Linux. It notes that HP has supported Linux on its servers for over 14 years through driver and firmware updates. It highlights some of HP's strategic partnerships with companies like Intel, Red Hat, and OpenStack. The document also summarizes tools and utilities HP provides to manage its servers running Linux, such as the Management Component Pack and Service Pack for ProLiant.
The document discusses HP's support for Linux on its ProLiant servers. It describes how HP tests commercial Linux distributions on ProLiant hardware, enables certification of distributions near new hardware launches, and provides the ProLiant Support Pack. It also discusses HP's Software Delivery Repository, which provides drivers, utilities, and tools. Finally, it explains HP tools like iLO, the Management Component Pack, and Service Pack for ProLiant, which provide additional monitoring, management, and update capabilities for ProLiant Linux systems.
In this deck from the 2016 HPC Advisory Council Switzerland Conference, Kenneth Hoste from the University Ghent presents an introduction to EasyBuild, an open-source framework for (automatically) getting scientific software installed on HPC systems.
Watch the video presentation: http://wp.me/p3RLHQ-f8J
Learn more: https://github.com/hpcugent/easybuild
See more talks from the Switzerland HPC Conference:
http://insidehpc.com/2016-swiss-hpc-conference/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Yocto Project Dev Day Prague 2017 - Advanced class - Kernel modules with eSDKMarco Cavallini
Excerpt of the lectures at Yocto Project Dev Day in Prague, 2017
During the advanced class Marco Cavallini ran this presentation about the creation of a Linux kernel module with eSDK.
OpenStack is an open source cloud computing platform that controls pools of compute, storage, and networking resources throughout a datacenter, managed through a dashboard that is exposed through APIs. It is made up of interrelated projects that handle functions like computing, networking, storage, imaging, orchestration, and more. The platform provides tools to provision resources to users in a simple and automated manner at scale.
OpenStack, Containers, and Docker: The Future of Application Deployment
Twenty years ago, developers built static applications on well-defined stacks that ran on proprietary, monolithic hardware. Developers today want freedom to build applications using their choice of services and stacks and, ideally, want to be able to run those applications on any available hardware. Of course, this raises questions about service interaction, the practicality of migrating applications across environments, and the challenges of managing unlimited combinations of services and hardware environment.
By promoting an opensource approach to flexible and inter-operable infrastructure, OpenStack goes a long way towards achieving this vision of the future. This talk discusses the application and platform side of the equation, and the interplay between OpenStack, Container technology (e.g. LXC), and the opensource Docker.io project. Docker.io enables any application and its dependencies to be deployed as lightweight containers that run consistently virtually anywhere. The same containerized application that runs on a developer's laptop can run consistently on a bare metal server, an OpenStack cluster, a Rackspace cloud, a VM,etc. While providing isolation and compatibility, containers have significant size, performance, and deployment advantages over traditional VMs.
Recently, the community created an integration between Docker and OpenStack Nova, opening up exciting possibilities for web scale application deployment, continuous integration and deployment, private PaaS, and hybrid cloud. This session will give an introduction to Docker and containers in the context of OpenStack, and will then demonstrate cross-environment deployment of applications.
This document discusses using Docker containers with OpenStack for application deployment. It begins with an introduction to Docker, describing its growth in usage and integration with various tools. Docker is presented as a solution to issues around deploying applications across different environments and hardware by providing lightweight, portable containers that package code and dependencies. The document demonstrates how Docker can be used with OpenStack through a new hypervisor that allows OpenStack to deploy and manage Linux containers, enabling control of Docker through the OpenStack dashboard. It outlines some benefits of Docker combined with OpenStack and the current state of the related OpenStack project.
Presentation given at AutoMacon on September 16, 2015 in Portland, OR. This talk covered how to build CoreOS components on Debian using Ansible and a brief demonstration of running an ELKStack application on the new cluster.
JDD2014: Docker.io - versioned linux containers for JVM devops - Dominik DornPROIDEA
This presentation will introduce you to Docker - the new shiny star on the Devops horizon. It will teach you everything you need to know to get started with Docker, why you'd want to use it and which tools to use to get the most out of it. Additionally to showing the basics, it will introduce helpful libraries available for the JVM and how they can be used together with Docker to create secure, scalable and maintainable cloud setups for your applications.
Performance characteristics of traditional v ms vs docker containers (dockerc...Boden Russell
Docker containers provide significantly lower resource usage and higher density than traditional virtual machines when running multiple workloads concurrently on a server.
When booting 15 Ubuntu VMs with MySQL sequentially, Docker containers boot on average 3.5 seconds compared to 5.8 seconds for KVMs. During steady state operation of 15 active VMs, Docker uses on average 0.2% CPU and 49MB RAM per container, while KVMs use 1.9% CPU and 292MB RAM each. Docker maintains low 1-minute load averages of 0.15, while KVMs average 35.9% under load.
Docker is an open platform for building and running distributed applications across multiple isolated containers. It allows encapsulating applications into lightweight executable packages called containers that include all necessary dependencies. The document discusses Docker's architecture, components, image layers, repositories, building images using Dockerfiles, and challenges in implementing Docker in real-world systems due to limitations in separating runtime configuration and data. Competition in the container space from projects like Rocket, Project Atomic, and Vagga is also discussed.
This talk gives a brief introduction to OpenStack and Chef, then outlines the current state of deploying OpenStack with Chef. There was a live demo deploying to a Dell rack during the talk.
SCALE 9x, February 25-27 in Los Angeles.
CoreOS automated MySQL Cluster Failover using Galera ClusterYazz Atlas
CoreOS Fleet and Etcd provide a simple and eloquent framework for application clusters to both auto-configure and recover from node failure. Galera Cluster is a multi-master, open solution for clustering MySQL. Mix the two, sprinkle in a bit of “glue” and you have a Docker based MySQL cluster that will react automatically to container failure. This presentation will cover the nuts and bolts of automating a Galera Cluster, built from Docker Images and deployed in a distributed fashion using etcd, confd, and fleet for both initial and failure recovery configuration.
LXC, Docker, and the future of software delivery | LinuxCon 2013dotCloud
This document discusses Linux containers and Docker. It describes how Linux containers provide isolation using namespaces and cgroups to allow applications to run consistently across different environments. Docker builds on Linux containers to make them easy to use, create, share, and deploy. Docker allows building images from Dockerfiles, sharing images in registries, and developing hybrid cloud workflows. The document outlines Docker's roadmap and growing ecosystem of tools and projects building on Docker.
Linux Container Brief for IEEE WG P2302Boden Russell
A brief into to Linux Containers presented to IEEE working group P2302 (InterCloud standards and portability). This deck covers:
- Definitions and motivations for containers
- Container technology stack
- Containers vs Hypervisor VMs
- Cgroups
- Namespaces
- Pivot root vs chroot
- Linux Container image basics
- Linux Container security topics
- Overview of Linux Container tooling functionality
- Thoughts on container portability and runtime configuration
- Container tooling in the industry
- Container gaps
- Sample use cases for traditional VMs
Overall, a bulk of this deck is covered in other material I have posted here. However there are a few new slides in this deck, most notability some thoughts on container portability and runtime config.
Redfish and python-redfish for Software Defined InfrastructureBruno Cornec
How the new Redfish protocol will help achieving the promises of a Software Defined Infrastructure, and which new projects are needed such as python-redfish and Alexandria to support it
Build and run applications in a dockerless kubernetes worldJorge Morales
Talk at Dev Days Riga 2018:
Kubernetes has rapidly grown to support many container runtime formats. In this talk, I'm presenting all the alternatives you have to run your applications in kubernetes, and will present CRI-O which is steadily becoming a replacement to run your Docker containers on production. And since you will no longer have Docker, how will you build now your Docker containers? Buildah is a project that facilitates building Docker containers in a Dockerless world.
Lxc – next gen virtualization for cloud intro (cloudexpo)Boden Russell
This document provides an introduction and overview of Linux containers as next-generation virtualization for cloud computing. It discusses how Linux containers provide better performance and flexibility than traditional virtual machines through the use of cgroups and namespaces. It also covers how containerization is gaining industry momentum and provides lower total cost of ownership through integration with modern Linux kernels and open source tooling. Finally, it defines key Linux container technologies, compares containers to hypervisors, and discusses building and securing Linux containers.
LXC – NextGen Virtualization for Cloud benefit realization (cloudexpo)Boden Russell
This document summarizes a presentation on Linux containers (LXC) as an alternative to virtual machines for cloud computing and containerization. It finds that LXC provides significantly better performance than virtual machines in terms of provisioning time, CPU and memory usage, and load on the compute node. Specifically, when packing 15 active VMs/containers on a node, LXC uses 0.54% CPU on average compared to 7.64% for KVM, and 734 MB of memory total compared to 4,387 MB for KVM. When booting VMs/containers serially, the average boot time is 3.5 seconds for LXC versus 5.8 seconds for KVM, and CPU usage is lower overall for
Isn’t it Ironic that a Redfish is software defining you Bruno Cornec
This document discusses combining the Ironic project, which provides bare-metal provisioning capabilities to OpenStack, with the Redfish standard developed by DMTF for RESTful management of server hardware. It proposes developing a Python Redfish library for Ironic to use, creating a Redfish driver for Ironic, and setting up a test environment using Docker containers to emulate a Redfish-based platform and validate the integration. Combining Ironic and Redfish could provide standard ways to manage hardware lifecycle actions and inventory systems through Redfish's REST API.
Presentation of the GPL Continuous Packaging tool helping producing packages mostly for upstream Open Source projects made during the Linux.conf.au 2014 Continuous Integration Mini-Conf in Perth.
OpenStack is an open source cloud computing platform that controls pools of compute, storage, and networking resources throughout a datacenter, managed through a dashboard that is exposed through APIs. It is made up of interrelated projects that handle functions like computing, networking, storage, imaging, orchestration, and more. The platform provides tools to provision resources to users in a simple and automated manner at scale.
OpenStack, Containers, and Docker: The Future of Application Deployment
Twenty years ago, developers built static applications on well-defined stacks that ran on proprietary, monolithic hardware. Developers today want freedom to build applications using their choice of services and stacks and, ideally, want to be able to run those applications on any available hardware. Of course, this raises questions about service interaction, the practicality of migrating applications across environments, and the challenges of managing unlimited combinations of services and hardware environment.
By promoting an opensource approach to flexible and inter-operable infrastructure, OpenStack goes a long way towards achieving this vision of the future. This talk discusses the application and platform side of the equation, and the interplay between OpenStack, Container technology (e.g. LXC), and the opensource Docker.io project. Docker.io enables any application and its dependencies to be deployed as lightweight containers that run consistently virtually anywhere. The same containerized application that runs on a developer's laptop can run consistently on a bare metal server, an OpenStack cluster, a Rackspace cloud, a VM,etc. While providing isolation and compatibility, containers have significant size, performance, and deployment advantages over traditional VMs.
Recently, the community created an integration between Docker and OpenStack Nova, opening up exciting possibilities for web scale application deployment, continuous integration and deployment, private PaaS, and hybrid cloud. This session will give an introduction to Docker and containers in the context of OpenStack, and will then demonstrate cross-environment deployment of applications.
This document discusses using Docker containers with OpenStack for application deployment. It begins with an introduction to Docker, describing its growth in usage and integration with various tools. Docker is presented as a solution to issues around deploying applications across different environments and hardware by providing lightweight, portable containers that package code and dependencies. The document demonstrates how Docker can be used with OpenStack through a new hypervisor that allows OpenStack to deploy and manage Linux containers, enabling control of Docker through the OpenStack dashboard. It outlines some benefits of Docker combined with OpenStack and the current state of the related OpenStack project.
Presentation given at AutoMacon on September 16, 2015 in Portland, OR. This talk covered how to build CoreOS components on Debian using Ansible and a brief demonstration of running an ELKStack application on the new cluster.
JDD2014: Docker.io - versioned linux containers for JVM devops - Dominik DornPROIDEA
This presentation will introduce you to Docker - the new shiny star on the Devops horizon. It will teach you everything you need to know to get started with Docker, why you'd want to use it and which tools to use to get the most out of it. Additionally to showing the basics, it will introduce helpful libraries available for the JVM and how they can be used together with Docker to create secure, scalable and maintainable cloud setups for your applications.
Performance characteristics of traditional v ms vs docker containers (dockerc...Boden Russell
Docker containers provide significantly lower resource usage and higher density than traditional virtual machines when running multiple workloads concurrently on a server.
When booting 15 Ubuntu VMs with MySQL sequentially, Docker containers boot on average 3.5 seconds compared to 5.8 seconds for KVMs. During steady state operation of 15 active VMs, Docker uses on average 0.2% CPU and 49MB RAM per container, while KVMs use 1.9% CPU and 292MB RAM each. Docker maintains low 1-minute load averages of 0.15, while KVMs average 35.9% under load.
Docker is an open platform for building and running distributed applications across multiple isolated containers. It allows encapsulating applications into lightweight executable packages called containers that include all necessary dependencies. The document discusses Docker's architecture, components, image layers, repositories, building images using Dockerfiles, and challenges in implementing Docker in real-world systems due to limitations in separating runtime configuration and data. Competition in the container space from projects like Rocket, Project Atomic, and Vagga is also discussed.
This talk gives a brief introduction to OpenStack and Chef, then outlines the current state of deploying OpenStack with Chef. There was a live demo deploying to a Dell rack during the talk.
SCALE 9x, February 25-27 in Los Angeles.
CoreOS automated MySQL Cluster Failover using Galera ClusterYazz Atlas
CoreOS Fleet and Etcd provide a simple and eloquent framework for application clusters to both auto-configure and recover from node failure. Galera Cluster is a multi-master, open solution for clustering MySQL. Mix the two, sprinkle in a bit of “glue” and you have a Docker based MySQL cluster that will react automatically to container failure. This presentation will cover the nuts and bolts of automating a Galera Cluster, built from Docker Images and deployed in a distributed fashion using etcd, confd, and fleet for both initial and failure recovery configuration.
LXC, Docker, and the future of software delivery | LinuxCon 2013dotCloud
This document discusses Linux containers and Docker. It describes how Linux containers provide isolation using namespaces and cgroups to allow applications to run consistently across different environments. Docker builds on Linux containers to make them easy to use, create, share, and deploy. Docker allows building images from Dockerfiles, sharing images in registries, and developing hybrid cloud workflows. The document outlines Docker's roadmap and growing ecosystem of tools and projects building on Docker.
Linux Container Brief for IEEE WG P2302Boden Russell
A brief into to Linux Containers presented to IEEE working group P2302 (InterCloud standards and portability). This deck covers:
- Definitions and motivations for containers
- Container technology stack
- Containers vs Hypervisor VMs
- Cgroups
- Namespaces
- Pivot root vs chroot
- Linux Container image basics
- Linux Container security topics
- Overview of Linux Container tooling functionality
- Thoughts on container portability and runtime configuration
- Container tooling in the industry
- Container gaps
- Sample use cases for traditional VMs
Overall, a bulk of this deck is covered in other material I have posted here. However there are a few new slides in this deck, most notability some thoughts on container portability and runtime config.
Redfish and python-redfish for Software Defined InfrastructureBruno Cornec
How the new Redfish protocol will help achieving the promises of a Software Defined Infrastructure, and which new projects are needed such as python-redfish and Alexandria to support it
Build and run applications in a dockerless kubernetes worldJorge Morales
Talk at Dev Days Riga 2018:
Kubernetes has rapidly grown to support many container runtime formats. In this talk, I'm presenting all the alternatives you have to run your applications in kubernetes, and will present CRI-O which is steadily becoming a replacement to run your Docker containers on production. And since you will no longer have Docker, how will you build now your Docker containers? Buildah is a project that facilitates building Docker containers in a Dockerless world.
Lxc – next gen virtualization for cloud intro (cloudexpo)Boden Russell
This document provides an introduction and overview of Linux containers as next-generation virtualization for cloud computing. It discusses how Linux containers provide better performance and flexibility than traditional virtual machines through the use of cgroups and namespaces. It also covers how containerization is gaining industry momentum and provides lower total cost of ownership through integration with modern Linux kernels and open source tooling. Finally, it defines key Linux container technologies, compares containers to hypervisors, and discusses building and securing Linux containers.
LXC – NextGen Virtualization for Cloud benefit realization (cloudexpo)Boden Russell
This document summarizes a presentation on Linux containers (LXC) as an alternative to virtual machines for cloud computing and containerization. It finds that LXC provides significantly better performance than virtual machines in terms of provisioning time, CPU and memory usage, and load on the compute node. Specifically, when packing 15 active VMs/containers on a node, LXC uses 0.54% CPU on average compared to 7.64% for KVM, and 734 MB of memory total compared to 4,387 MB for KVM. When booting VMs/containers serially, the average boot time is 3.5 seconds for LXC versus 5.8 seconds for KVM, and CPU usage is lower overall for
Isn’t it Ironic that a Redfish is software defining you Bruno Cornec
This document discusses combining the Ironic project, which provides bare-metal provisioning capabilities to OpenStack, with the Redfish standard developed by DMTF for RESTful management of server hardware. It proposes developing a Python Redfish library for Ironic to use, creating a Redfish driver for Ironic, and setting up a test environment using Docker containers to emulate a Redfish-based platform and validate the integration. Combining Ironic and Redfish could provide standard ways to manage hardware lifecycle actions and inventory systems through Redfish's REST API.
Presentation of the GPL Continuous Packaging tool helping producing packages mostly for upstream Open Source projects made during the Linux.conf.au 2014 Continuous Integration Mini-Conf in Perth.
The document provides an overview of the Docker ecosystem, including its definition, architecture, and status. It describes how Docker allows for applications to be bundled and run in a portable way across various environments using containers. The key components of Docker like images, containers, registries, and Dockerfiles are explained. The document also discusses the container ecosystem and adoption of Docker by various companies and projects. It outlines the security features and best practices for containers. Finally, it provides a brief history of resource management capabilities in Linux that enabled and influenced the development of containers.
Continuous Packaging is also Mandatory for DevOpsDocker, Inc.
While DevOps are comfortable with continuous integration and automatic tests, the area of continuous packaging has not been given the attention it deserves.
Even with containers, delivering an application using software packages provides multiple advantages with regards to file-based installation: it allows to manage dependencies more easily, to provide metadata, checksum, and signature mechanisms, to deal with packages repositories.
But doing that in a continuous packaging approach means that the generation of these packages is fully automated and part of the build process of the software. As a consequence, it eases the various steps of a solution lifecycle (controlled impact of installation/uninstallation,
identical deliveries up to the customer, avoidance of code or metadata duplication)
This presentation will detail the methodological approach around continuous packaging and demonstrate how this can be put in place using an Open Source tool such as project-builder.org and how this allows the MondoRescue project to deliver packages at will for lots of distribution tuples through the same number of Docker containers.
This document provides information about Linux containers and Docker. It discusses:
1) The evolution of IT from client-server models to thin apps running on any infrastructure and the challenges of ensuring consistent service interactions and deployments across environments.
2) Virtual machines and their benefits of full isolation but large disk usage, and Vagrant which allows packaging and provisioning of VMs via files.
3) Docker and how it uses Linux containers powered by namespaces and cgroups to deploy applications in lightweight portable containers that are more efficient than VMs. Examples of using Docker are provided.
Docker - Demo on PHP Application deployment Arun prasath
Docker is an open-source project to easily create lightweight, portable, self-sufficient containers from any application. The same container that a developer builds and tests on a laptop can run at scale, in production, on VMs, bare metal, OpenStack clusters, public clouds and more.
In this demo, I will show how to build a Apache image from a Dockerfile and deploy a PHP application which is present in an external folder using custom configuration files.
This document discusses the latest trends for cloud native application development on OpenShift 4. It covers OpenShift's focus on simplifying creation of cloud native services and serverless functions using components and tools without requiring deep Kubernetes knowledge. Developer tools like CodeReady Workspaces and the odo CLI aim to improve developer productivity. Operators are highlighted as a way to automate application management. Knative and service mesh technologies are discussed as ways to enable event-driven and microservices-based applications. OpenShift 4's new installation process and ability to perform over-the-air updates are also summarized.
Docker - A high level introduction to dockers and containersDr Ganesh Iyer
A high level introduction to Dockers and Containers. Many of the slides are not mine.I used the slides I got from Internet and prepared the rest of the slides based on my understand form various blogs and other google info.
This document discusses using MySQL containers and provides an overview of containers and Docker. It begins with definitions of containers and how they isolate applications. It then discusses Docker and how it packages and runs applications in containers. The rest of the document demonstrates how to use official MySQL containers, including running MySQL in containers and using Docker Compose to manage multi-container applications. It concludes by discussing container orchestration with Docker Swarm and Kubernetes.
State of Containers and the Convergence of HPC and BigDatainside-BigData.com
In this deck from 2018 Swiss HPC Conference, Christian Kniep from Docker Inc. presents: State of Containers and the Convergence of HPC and BigData.
"This talk will recap the history of and what constitutes Linux Containers, before laying out how the technology is employed by various engines and what problems these engines have to solve. Afterward Christian will elaborate on why the advent of standards for images and runtimes moved the discussion from building and distributing containers to orchestrating containerized applications at scale. In conclusion attendees will get an update on how containers foster the convergence of Big Data and HPC workloads and the state of native HPC containers."
Learn more: http://docker.com
and
http://www.hpcadvisorycouncil.com/events/2018/swiss-workshop/agenda.php
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
The document discusses OpenStack and Fibre Channel storage. It provides an overview of OpenStack, including its goals of being an open platform with broad support and empowering users. It describes core OpenStack technologies like Compute, Object Storage, and Block Storage. It outlines the history and current state of Fibre Channel support in OpenStack, including the Fibre Channel Zone Manager that automates zoning. It diagrams the high-level architecture and components involved in provisioning Fibre Channel volumes to virtual machines from OpenStack.
The Information Technology have led us into an era where the production, sharing and use of information are now part of everyday life and of which we are often unaware actors almost: it is now almost inevitable not leave a digital trail of many of the actions we do every day; for example, by digital content such as photos, videos, blog posts and everything that revolves around the social networks (Facebook and Twitter in particular). Added to this is that with the "internet of things", we see an increase in devices such as watches, bracelets, thermostats and many other items that are able to connect to the network and therefore generate large data streams. This explosion of data justifies the birth, in the world of the term Big Data: it indicates the data produced in large quantities, with remarkable speed and in different formats, which requires processing technologies and resources that go far beyond the conventional systems management and storage of data. It is immediately clear that, 1) models of data storage based on the relational model, and 2) processing systems based on stored procedures and computations on grids are not applicable in these contexts. As regards the point 1, the RDBMS, widely used for a great variety of applications, have some problems when the amount of data grows beyond certain limits. The scalability and cost of implementation are only a part of the disadvantages: very often, in fact, when there is opposite to the management of big data, also the variability, or the lack of a fixed structure, represents a significant problem. This has given a boost to the development of the NoSQL database. The website NoSQL Databases defines NoSQL databases such as "Next Generation Databases mostly addressing some of the points: being non-relational, distributed, open source and horizontally scalable." These databases are: distributed, open source, scalable horizontally, without a predetermined pattern (key-value, column-oriented, document-based and graph-based), easily replicable, devoid of the ACID and can handle large amounts of data. These databases are integrated or integrated with processing tools based on the MapReduce paradigm proposed by Google in 2009. MapReduce with the open source Hadoop framework represent the new model for distributed processing of large amounts of data that goes to supplant techniques based on stored procedures and computational grids (step 2). The relational model taught courses in basic database design, has many limitations compared to the demands posed by new applications based on Big Data and NoSQL databases that use to store data and MapReduce to process large amounts of data.
Course Website http://pbdmng.datatoknowledge.it/
Contact me to download the slides
My Galera on Kubernetes on CoreOS presentation from Percona Live 2015 in Santa Clara. Please be patient as I need to edit my videos and upload them to youtube in the next few days.
Linux Containers and Docker SHARE.ORG Seattle 2015Filipe Miranda
This slide deck shows us an introduction to Linux Containers (LXC) and Docker for Linux on IBM z Systems.
One example of a commercial use of Linux Containers (and Docker) is Red Hat Openshift, which is is also covered at the end.
Docker allows building and running applications inside lightweight containers. Some key benefits of Docker include:
- Portability - Dockerized applications are completely portable and can run on any infrastructure from development machines to production servers.
- Consistency - Docker ensures that application dependencies and environments are always the same, regardless of where the application is run.
- Efficiency - Docker containers are lightweight since they don't need virtualization layers like VMs. This allows for higher density and more efficient use of resources.
This two-day training covers Docker concepts including installation, working with containers and images, building images with Dockerfiles, and integrating Docker with OpenStack. Day one focuses on the Docker introduction, installation, containers, images, Dockerfiles, and using Nova to run Docker containers as compute instances. It also covers using Glance as a Docker image registry. Day two covers Docker clustering with Kubernetes, networking, Docker Hub, case studies, and the Docker source code. It concludes with developing platforms and running Hadoop on Docker containers.
Deploying an Open Source DAM in SAAS Mode (European Drupal Days 2015)Eugenio Minardi
This document summarizes Michel van de Ven's presentation on deploying the open source digital asset management system MediaMosa in a software as a service (SAAS) model. MediaMosa is a Drupal-based system that allows users to store, manage, and deliver digital media and assets. Inuits deploys MediaMosa instances for its clients using automation tools like Puppet, Jenkins, and Mcollective to manage infrastructure as code and implement continuous delivery pipelines. This allows Inuits to efficiently operate MediaMosa in a distributed and scalable SAAS environment.
Red Hat OpenShift 4 allows for automated and customized deployments. The Full Stack Automation method fully automates installation and updates of both the OpenShift platform and Red Hat Enterprise Linux CoreOS host operating system. The Pre-existing Infrastructure method allows OpenShift to be deployed on user-managed infrastructure, where the customer provisions resources like load balancers and DNS. Both methods use the openshift-install tool to generate ignition configs and monitor the cluster deployment.
This document discusses HP Helion OpenStack, which is HP's distribution of the OpenStack cloud computing platform. It provides an overview of OpenStack and HP's leadership within the OpenStack community as a major contributor. HP Helion OpenStack adds enterprise-grade features to OpenStack such as security, high availability, and support. It allows organizations to build both private and public clouds using an open source technology.
The document discusses Red Hat OpenShift 4 installation methods. It describes the Full Stack Automation method where the installer provisions all infrastructure components including hosts running Red Hat Enterprise Linux CoreOS (RHCOS). It also covers deploying to pre-existing infrastructure where the user provisions infrastructure like VMs, load balancers, and DNS, while the installer configures the OpenShift cluster and RHCOS hosts.
Semelhante a Multi-OS Continuous Packaging with docker and Project-Builder.org (20)
Diaporama réalisé pour les Concerts 2022 de l'ensemble vocal Variations pour le 501è anniversaire de la mort de Josquin Desprez.
Pièces chantées:
O salutaris hostia
Tu solus
Stabat Mater dolorosa
Scaramella
Ballade des pendus
Miserere mei Deus
Cœurs désolés
Sanctus de la messe L’Homme Armé
Christus mortuus est
Laudate pueri Dominum
El grillo
Proch Dolor
Mille regretz
Tulerunt Dominum meum
Redfish is an IPMI replacement standardized by the DMTF. It provides a RESTful API for server out of band management and a lightweight data model specification that is scalable, discoverable and extensible. (Cf: http://www.dmtf.org/standards/redfish). This presentation will start by detailing its role and the features it provides with examples. It will demonstrate the benefits it provides to system administrator by providing a standardized open interface for multiple servers, and also storage systems.
We will then cover various tools such as the DMTF ones and the python-redfish library (Cf: https://github.com/openstack/python-redfish) offering Redfish abstractions.
Document issu d'un groupe de travail du Syntec auquel j'ai participé et qui n'est plus accessible via les sites originaux donc je mets une copie ici.
Il s'agit d'un document partageant des bonnes pratiques autour de la gouvernance et du pilotage de projets Open Source dans des organismes.
Using containers and Continuous Packaging to Build native FOSSology packagesBruno Cornec
During last LinuxCon, Bruno presented the continuous packaging approach used with a tool like project-builder.org to package upstream projects for hundreds of Linux distributions tuples in an automatic manner. Discussions happened there with the FOSSology project which wanted to benefit from this approach to produce Linux packages for their users. Both projects have since that worked jointly to make it a reality, and want to share their return of experience on this journey, benefits obtained, issues encountered and how they were fixed.
After a reminder of the basics on continuous packing, the presentation will give a concrete example of what was setup using the infrastructure of the LinuxFoudation to enable the automatic creation of rpm and deb packages for FOSSology, launched during the continuous integration process already in place. A demo of the build process will also be made.
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
Digital Banking in the Cloud: How Citizens Bank Unlocked Their MainframePrecisely
Inconsistent user experience and siloed data, high costs, and changing customer expectations – Citizens Bank was experiencing these challenges while it was attempting to deliver a superior digital banking experience for its clients. Its core banking applications run on the mainframe and Citizens was using legacy utilities to get the critical mainframe data to feed customer-facing channels, like call centers, web, and mobile. Ultimately, this led to higher operating costs (MIPS), delayed response times, and longer time to market.
Ever-changing customer expectations demand more modern digital experiences, and the bank needed to find a solution that could provide real-time data to its customer channels with low latency and operating costs. Join this session to learn how Citizens is leveraging Precisely to replicate mainframe data to its customer channels and deliver on their “modern digital bank” experiences.
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsDianaGray10
Join us to learn how UiPath Apps can directly and easily interact with prebuilt connectors via Integration Service--including Salesforce, ServiceNow, Open GenAI, and more.
The best part is you can achieve this without building a custom workflow! Say goodbye to the hassle of using separate automations to call APIs. By seamlessly integrating within App Studio, you can now easily streamline your workflow, while gaining direct access to our Connector Catalog of popular applications.
We’ll discuss and demo the benefits of UiPath Apps and connectors including:
Creating a compelling user experience for any software, without the limitations of APIs.
Accelerating the app creation process, saving time and effort
Enjoying high-performance CRUD (create, read, update, delete) operations, for
seamless data management.
Speakers:
Russell Alfeche, Technology Leader, RPA at qBotic and UiPath MVP
Charlie Greenberg, host
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
What is an RPA CoE? Session 1 – CoE VisionDianaGray10
In the first session, we will review the organization's vision and how this has an impact on the COE Structure.
Topics covered:
• The role of a steering committee
• How do the organization’s priorities determine CoE Structure?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
"Choosing proper type of scaling", Olena SyrotaFwdays
Imagine an IoT processing system that is already quite mature and production-ready and for which client coverage is growing and scaling and performance aspects are life and death questions. The system has Redis, MongoDB, and stream processing based on ksqldb. In this talk, firstly, we will analyze scaling approaches and then select the proper ones for our system.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
AppSec PNW: Android and iOS Application Security with MobSFAjin Abraham
Mobile Security Framework - MobSF is a free and open source automated mobile application security testing environment designed to help security engineers, researchers, developers, and penetration testers to identify security vulnerabilities, malicious behaviours and privacy concerns in mobile applications using static and dynamic analysis. It supports all the popular mobile application binaries and source code formats built for Android and iOS devices. In addition to automated security assessment, it also offers an interactive testing environment to build and execute scenario based test/fuzz cases against the application.
This talk covers:
Using MobSF for static analysis of mobile applications.
Interactive dynamic security assessment of Android and iOS applications.
Solving Mobile app CTF challenges.
Reverse engineering and runtime analysis of Mobile malware.
How to shift left and integrate MobSF/mobsfscan SAST and DAST in your build pipeline.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
In the realm of cybersecurity, offensive security practices act as a critical shield. By simulating real-world attacks in a controlled environment, these techniques expose vulnerabilities before malicious actors can exploit them. This proactive approach allows manufacturers to identify and fix weaknesses, significantly enhancing system security.
This presentation delves into the development of a system designed to mimic Galileo's Open Service signal using software-defined radio (SDR) technology. We'll begin with a foundational overview of both Global Navigation Satellite Systems (GNSS) and the intricacies of digital signal processing.
The presentation culminates in a live demonstration. We'll showcase the manipulation of Galileo's Open Service pilot signal, simulating an attack on various software and hardware systems. This practical demonstration serves to highlight the potential consequences of unaddressed vulnerabilities, emphasizing the importance of offensive security practices in safeguarding critical infrastructure.