logo
Dates

Author


Conferences

Tags

Sort by:  

Authors: Todd Ekenstam, Phillip Wittrock, Aparna Subramanian, Nagarajan Chinnakaveti Thulasiraman
2023-04-21

Have you considered recession-proofing your platform? What steps does one need to take to reduce cost and increase efficiency? Join Shopify, Intuit, Apple, and Zalando at this panel discussion, where they share platform efficiency and cost optimization strategies. Learn about why this is a business priority, how to drive costs down when driving efficiency up, and what you can do to build efficient apps. You will walk away with concrete strategies such as defining measurable metrics, choosing the best autoscaling approach, and learning about instrumenting your platform for cost visibility and efficiency. Learn how to approach this complex problem with solutions that have worked for other end users. Be ready for any impending gloom or boom!
Authors: Sanjay Pujare, Costin Manolache
2023-04-21

Kubernetes is suitable for stateless services and one of its benefits is seamless autoscaling of infrastructure in response to varying load. The resulting elasticity enables users to optimize their infrastructure and minimize costs. But what do you do if your application is stateful - for example it requires maintaining stateful sessions between clients and servers and you are using a service mesh? In this talk we will cover stateful applications where clients can create and maintain persistent sessions in the presence of load balancing where Istio routes individual RPCs to various backends. Client creates a persistent session and expects all RPCs in that session to go to a particular backend because the backend has the “state” related to that session and is not able to share the state with other backends. Note that consistent hashing or rendezvous load balancing don’t quite work because session persistence is broken when the set of backends changes. The feature uses HTTP Cookies where persistent or stateful sessions are achieved by communicating with the load balancer via cookies in Istio and proxyless gRPC. We also cover the use-case of “draining backends” where backends are gradually removed as part of downsizing the infrastructure but without breaking session persistence.
Authors: Zbynek Roubalik, Jorge Turrado
2023-04-20

tldr - powered by Generative AI

The presentation discusses the importance of certificate management and web hook validation in Keda, a Kubernetes-based event-driven autoscaler.
  • Encrypting internal traffic inside the cluster is necessary to prevent unauthorized access and scaling issues
  • Keda introduces mechanisms for automatically generating TLS certificates and supports the use of custom CA
  • Validation webhooks prevent scaling conflicts and ensure that required metrics are present
  • Managed identities are a secure way to connect to cloud provider infrastructure
  • Exposing metrics is critical for monitoring Keda's performance
Authors: Marcin Wielgus, Diego Bonfigli, Jayant Jain
2022-10-28

tldr - powered by Generative AI

Updates and features of Kubernetes Autoscaling community
  • Updates on Horizontal Pod Autoscaler v2 API and deprecation notices for v2 beta1 and v2 beta2
  • Introduction of gRPC editions in Cluster Autoscaler, including the grpc expander and provider
  • Changes in Vertical Pod Autoscaler
  • Encouragement for community involvement in bug reports, pull requests, and documentation improvement
Authors: Michael McCune, Guy Templeton, David Morrison, Joachim Bartosik
2022-05-19

tldr - powered by Generative AI

Updates and features from the Kubernetes Autoscaling community
  • Horizontal Pod Auto Scaler has released a v2 API and deprecated v2 beta1 and v2 beta2
  • No changes to the serialization format
  • Programmatic API interface has changed
  • Cluster Auto Scaler has added gRPC extensions
  • Vertical Pod Auto Scaler has undergone changes
  • Community involvement is encouraged
Authors: Ciprian Hacman, Radu Gheorghe
2022-05-19

tldr - powered by Generative AI

Best practices for scaling Elasticsearch clusters
  • Use metrics from inside Elasticsearch for accuracy
  • Scale in larger increments to reduce noise
  • Force index rotation to evenly spread load across nodes
  • Judge cluster size based on disk usage and search latency
  • Use local SSDs for better I/O latency
  • Consider hot-warm-cold architecture for data management
Authors: Bowen Li, huichao zhao
2022-05-18

tldr - powered by Generative AI

The presentation discusses the design principles and architecture of a cloud-native Spark on Kubernetes platform, highlighting the benefits of cloud and Kubernetes and the need for auto-scaling based on cost-saving and elasticity.
  • Cloud and Kubernetes can solve problems of legacy infrastructure by providing on-demand, elastic, and scalable resources with strong resource isolation and cutting-edge security techniques.
  • Design principles include fully embracing public cloud and cognitive way of thinking, containerization for elasticity and reproducibility, and decoupling compute and storage for independent scaling.
  • The architecture of the cloud-native Spark on Kubernetes platform involves multiple Spark Kubernetes clusters, a Spark service gateway, and a multi-tenant platform with advanced features such as physical isolation and min/max capacity setting.
  • Auto-scaling is necessary for cost-saving and elasticity, and the presentation discusses the design of reactive auto-scaling and its productionization.
  • The platform has been running in production for a year, supporting many business-critical workloads for Apple AML.
Authors: Natalie Serrino
2022-05-18

tldr - powered by Generative AI

Autoscaling Kubernetes Deployments is a flexible and rich option for ensuring stable performance when the load on the application changes over time.
  • Factors to consider when sizing your Kubernetes application
  • Horizontal vs Vertical autoscaling
  • Selecting the right auto scaling metric for your application
  • A Turing-complete autoscaler demo
Authors: Marcin Wielgus, Joseph Burnett
2021-10-14

tldr - powered by Generative AI

An overview of autoscaling features in Kubernetes, covering horizontal, vertical, and cluster autoscalers, and upcoming features like the new HPA v2 stable API and support for alternative recommenders in VPA.
  • SIG Auto Scaling is responsible for horizontal and vertical polynomial scaling, cluster proportional system component auto scaling, and cluster autoscaler.
  • Horizontal Pod Autoscaler (HPA) scales in and out based on pre-defined metrics, while Vertical Pod Autoscaler (VPA) adjusts resources for workloads that don't necessarily scale horizontally.
  • Cluster autoscaler adjusts the number of nodes in any of the node pools that are configured.
  • The new HPA v2 stable API and support for alternative recommenders in VPA are upcoming features.
  • The autoscaling community can be accessed through Slack channels and weekly meetings.
Authors: Mulugeta Ayalew Tamiru
2021-10-13

tldr - powered by Generative AI

Mck8s is a container orchestration platform for geo-distributed multi-cluster environments that aims to bring placement, autoscaling, bursting, inter-cluster routing, and cluster provisioning into one package.
  • Cloud deployments have become increasingly geographically distributed, leading to challenges in managing resources and deploying applications
  • Mck8s addresses these challenges by providing automated ways of deploying applications and managing resources, including solving problems related to resilience, scaling, user traffic routing, and load balancing
  • Mck8s is built on top of Kubernetes and offers a more integrated and autonomous approach for managing geo-distributed Kubernetes clusters at scale
  • Mck8s emphasizes usability and easy adoption by using manifest files similar to vanilla Kubernetes