logo
Dates

Author


Conferences

Tags

Sort by:  

Authors: Felix Hoffmann
2023-04-19

tldr - powered by Generative AI

The presentation discusses resource management in Kubernetes from the perspective of an application developer, highlighting the importance of setting resource requests and limits appropriately to avoid cluster crashes and scheduling issues.
  • Resource management in Kubernetes involves setting CPU and memory requests and limits for containers
  • Memory limits result in termination of pods when exceeded, while CPU limits can lead to throttling or termination
  • Setting appropriate requests and limits is crucial for efficient scheduling and avoiding noisy neighbors
  • Developers should be aware of namespace limits and available resources when setting requests and limits
  • In general, it is advisable to set memory requests equal to memory limits and avoid setting CPU limits
  • Exceptions include cases where consistent workloads or overcommitment of memory are preferred
Conference:  ContainerCon 2022
Authors: Ying Xiong, Yulin Sun
2022-06-24

Quark container is one new secure container runtime as Kata runtime and gVisor. It is a high-performance OCI compatible secure container runtime developed by Rust program language with low memory footprint and startup overhead. It is designed for containerized service execution running in the Data Center and Edge environment. Quark is based on System Call virtualization. Compare with Linux VM based virtualization, it adopts multiple performance improvements such as zero memory copy QCall to avoid Hypercall’s communication overhead between Guest Kernel and Host; Linux host io-uring based IO virtualization. It also support RDMA based container network which transparently execute TCP based containerized application over RDMA to achieve low latency and high throughput. The talk will introduce Quark container architecture together with the secure container design challenge and opportunity.
Conference:  ContainerCon 2022
Authors: Phu Tran, Vinay Kulkarni
2022-06-23

tldr - powered by Generative AI

The presentation discusses the use of ebpf technology in achieving cni networking with Mizar and XDP. The speaker also talks about future plans for enhancing the technology and proposes a formal EPA change to Kubernetes.
  • Ebpf technology was used to achieve cni networking with Mizar and XDP without changing any lines of kernel code
  • Future plans include enhancing the technology with a tx hook for XDP, proposing a formal EPA change to Kubernetes, and improving performance measurement
  • The speaker also discusses the need for a management plane and multi-tenant networking
  • The presentation includes a demo of the technology using four virtual machines
Authors: Bowen Li, huichao zhao
2022-05-18

tldr - powered by Generative AI

The presentation discusses the design principles and architecture of a cloud-native Spark on Kubernetes platform, highlighting the benefits of cloud and Kubernetes and the need for auto-scaling based on cost-saving and elasticity.
  • Cloud and Kubernetes can solve problems of legacy infrastructure by providing on-demand, elastic, and scalable resources with strong resource isolation and cutting-edge security techniques.
  • Design principles include fully embracing public cloud and cognitive way of thinking, containerization for elasticity and reproducibility, and decoupling compute and storage for independent scaling.
  • The architecture of the cloud-native Spark on Kubernetes platform involves multiple Spark Kubernetes clusters, a Spark service gateway, and a multi-tenant platform with advanced features such as physical isolation and min/max capacity setting.
  • Auto-scaling is necessary for cost-saving and elasticity, and the presentation discusses the design of reactive auto-scaling and its productionization.
  • The platform has been running in production for a year, supporting many business-critical workloads for Apple AML.
Authors: Mauro Pessina
2022-05-18

tldr - powered by Generative AI

The presentation discusses an AI-powered optimization methodology for improving cost efficiency and performance of digital services provided by a company.
  • The challenge faced by the customer was to optimize their application while keeping on releasing application updates to introduce new business functionalities and align to new regulations.
  • The tuning practice in place was manual and took almost two months to tune one single macro service.
  • The AI-powered optimization methodology works in five steps: applying new configuration suggested by AI, applying workload to target system, collecting KPIs, analyzing results, and producing new configuration to be tested in the next iteration.
  • The methodology allows setting constraints and goals, such as minimizing application cost and ensuring service reliability.
  • The presentation provides an anecdote of how the methodology was used to optimize a customer's authentication service on Kubernetes, resulting in a 49% improvement on cost efficiency compared to the baseline configuration.
Authors: Natalie Serrino
2022-05-18

tldr - powered by Generative AI

Autoscaling Kubernetes Deployments is a flexible and rich option for ensuring stable performance when the load on the application changes over time.
  • Factors to consider when sizing your Kubernetes application
  • Horizontal vs Vertical autoscaling
  • Selecting the right auto scaling metric for your application
  • A Turing-complete autoscaler demo
Authors: Madalina Lazar, Denisio Togashi
2022-05-18

tldr - powered by Generative AI

Telemetry Aware Scheduling is an open-source project that uses telemetry to make smarter scheduling decisions for workloads in Kubernetes clusters.
  • Telemetry Aware Scheduling (TAS) is an open-source project that extends Kubernetes' scheduling paradigm to use knowledge of resources to impact scheduling decisions.
  • TAS uses telemetry to help make scheduling decisions and is an extender of the Kubernetes scheduler.
  • TAS allows for filtering and scoring nodes and utilizes node affinity rules via fixed and custom labels.
  • TAS uses telemetry where scheduling policies that are structurally based on rules which are based on metrics that come from the cluster.
  • TAS requires a metrics pipeline to expose, collect, store, and make metrics available to the Kubernetes custom metrics API.
  • TAS works together with the default scheduler and returns a suggested outcome of pod placement to the default scheduler.
  • TAS supports multi-metric rules that contain multiple metrics and can link them together with operators such as any off or all of.
Authors: Kemal Akkoyun, Matthias Loibl
2021-10-15

tldr - powered by Generative AI

The presentation discusses the development of Parka, a tool for profiling and analyzing performance in software applications. The focus is on the storage architecture and the process of handling write requests.
  • Parka is a tool for profiling and analyzing performance in software applications
  • The storage architecture of Parka is designed to handle stack traces as a first-class citizen
  • Write requests are ingested and validated using protobuf and metadata label sets
  • The metadata store is implemented in SQLite and can be used with any SQL database
  • The end result of a write request is a set of location IDs and corresponding sample values