Solution guide: Archive your cold data to Google Cloud Storage with Komprise

More than 56% of enterprises have more than half a petabyte of inactive data but this “cold” data often lives on expensive primary storage platforms. Google Cloud Storage provides an opportunity to store this data cost-effectively and achieve significant savings, but storage and IT admins often face the challenge of how to identify cold data and move it non-disruptively.

Komprise, a Google Cloud technology partner, provides software that analyzes data across NFS and SMB/CIFS storage to identify inactive/cold data, and moves the data transparently to Cloud Storage, which can help to cut costs significantly. Working with Komprise, we’ve prepared a full tutorial guide that describes how customers can understand data usage and growth in their storage environment, get a customized ROI analysis and move this data to Cloud Storage based on specific policies.

Cloud Storage provides excellent options to customers looking to store infrequently accessed data at low cost using Nearline or Coldline storage tiers. If and when access to this data is needed, there are no access time penalties; the data is available almost immediately. In addition, built-in object-level lifecycle management in Cloud Storage reduces the burden for admins by enabling policy-based movement of data across storage classes. With Komprise, customers can bring lifecycle management to their on-premise primary storage platforms and seamlessly move this data to the Cloud. Komprise deploys in under 15 minutes, works across NFS, SMB/CIFS and object storage without any storage agents, adapts to file-system and network loads to run non-intrusively in the background and scales out on-demand.

Teams can get started through this self-service tutorial or watch this on-demand webinar featuring Komprise’ COO Krishna Subramanian and Google Cloud Storage Product Manager Ben Chong. As always, don’t hesitate to reach out to us to explore which enterprise workloads make the most sense for your cloud initiatives.
Quelle: Google Cloud Platform

Scalability updates in Kubernetes 1.6: 5,000 node and 150,000 pod clusters

Editor’s note: this post is part of a series of in-depth articles on what’s new in Kubernetes 1.6Last summer we shared updates on Kubernetes scalability, since then we’ve been working hard and are proud to announce that Kubernetes 1.6 can handle 5,000-node clusters with up to 150,000 pods. Moreover, those cluster have even better end-to-end pod startup time than the previous 2,000-node clusters in the 1.3 release; and latency of the API calls are within the one-second SLO.In this blog post we review what metrics we monitor in our tests and describe our performance results from Kubernetes 1.6. We also discuss what changes we made to achieve the improvements, and our plans for upcoming releases in the area of system scalability.X-node clusters – what does it mean?Now that Kubernetes 1.6 is released, it is a good time to review what it means when we say we “support” X-node clusters. As described in detail in a previous blog post, we currently have two performance-related Service Level Objectives (SLO):API-responsiveness: 99% of all API calls return in less than 1sPod startup time: 99% of pods and their containers (with pre-pulled images) start within 5s.As before, it is possible to run larger deployments than the stated supported 5,000-node cluster (and users have), but performance may be degraded and it may not meet our strict SLO defined above.We are aware of the limited scope of these SLOs. There are many aspects of the system that they do not exercise. For example, we do not measure how soon a new pod that is part of a service will be reachable through the service IP address after the pod is started. If you are considering using large Kubernetes clusters and have performance requirements not covered by our SLOs, please contact the Kubernetes Scalability SIG so we can help you understand whether Kubernetes is ready to handle your workload now.The top scalability-related priority for upcoming Kubernetes releases is to enhance our definition of what it means to support X-node clusters by:refining currently existing SLOsadding more SLOs (that will cover various areas of Kubernetes, including networking)Kubernetes 1.6 performance metrics at scaleSo how does performance in large clusters look in Kubernetes 1.6? The following graph shows the end-to-end pod startup latency with 2000- and 5000-node clusters. For comparison, we also show the same metric from Kubernetes 1.3, which we published in our previous scalability blog post that described support for 2000-node clusters. As you can see, Kubernetes 1.6 has better pod startup latency with both 2000 and 5000 nodes compared to Kubernetes 1.3 with 2000 nodes [1].The next graph shows API response latency for a 5000-node Kubernetes 1.6 cluster. The latencies at all percentiles are less than 500ms, and even 90th percentile is less than about 100ms.How did we get here?Over the past nine months (since the last scalability blog post), there have been a huge number of performance and scalability related changes in Kubernetes. In this post we will focus on the two biggest ones and will briefly enumerate a few others.etcd v3In Kubernetes 1.6 we switched the default storage backend (key-value store where the whole cluster state is stored) from etcd v2 to etcd v3. The initial works towards this transition has been started during the 1.3 release cycle. You might wonder why it took us so long, given that:the first stable version of etcd supporting the v3 API was announced on June 30, 2016the new API was designed together with the Kubernetes team to support our needs (from both a feature and scalability perspective)the integration of etcd v3 with Kubernetes had already mostly been finished when etcd v3 was announced (indeed CoreOS used Kubernetes as a proof-of-concept for the new etcd v3 API)As it turns out, there were a lot of reasons. We will describe the most important ones below.Changing storage in a backward incompatible way, as is in the case for the etcd v2 to v3 migration, is a big change, and thus one for which we needed a strong justification. We found this justification in September when we determined that we would not be able to scale to 5000-node clusters if we continued to use etcd v2 (kubernetes/32361 contains some discussion about it). In particular, what didn’t scale was the watch implementation in etcd v2. In a 5000-node cluster, we need to be able to send at least 500 watch events per second to a single watcher, which wasn’t possible in etcd v2.Once we had the strong incentive to actually update to etcd v3, we started thoroughly testing it. As you might expect, we found some issues. There were some minor bugs in Kubernetes, and in addition we requested a performance improvement in etcd v3’s watch implementation (watch was the main bottleneck in etcd v2 for us). This led to the 3.0.10 etcd patch release.Once those changes had been made, we were convinced that new Kubernetes clusters would work with etcd v3. But the large challenge of migrating existing clusters remained. For this we needed to automate the migration process, thoroughly test the underlying CoreOS etcd upgrade tool, and figure out a contingency plan for rolling back from v3 to v2.But finally, we are confident that it should work.Switching storage data format to protobufIn the Kubernetes 1.3 release, we enabled protobufs as the data format for Kubernetes components to communicate with the API server (in addition to maintaining support for JSON). This gave us a huge performance improvement.However, we were still using JSON as a format in which data was stored in etcd, even though technically we were ready to change that. The reason for delaying this migration was related to our plans to migrate to etcd v3. Now you are probably wondering how this change was depending on migration to etcd v3. The reason for it was that with etcd v2 we couldn’t really store data in binary format (to workaround it we were additionally base64-encoding the data), whereas with etcd v3 it just worked. So to simplify the transition to etcd v3 and avoid some non-trivial transformation of data stored in etcd during it, we decided to wait with switching storage data format to protobufs until migration to etcd v3 storage backend is done.Other optimizationsWe made tens of optimizations throughout the Kubernetes codebase during the last three releases, including:optimizing the scheduler (which resulted in 5-10x higher scheduling throughput)switching all controllers to a new recommended design using shared informers, which reduced resource consumption of controller-manager – for reference see this documentoptimizing individual operations in the API server (conversions, deep-copies, patch)reducing memory allocation in the API server (which significantly impacts the latency of API calls)We want to emphasize that the optimization work we have done during the last few releases, and indeed throughout the history of the project, is a joint effort by many different companies and individuals from the whole Kubernetes community.What’s next?People frequently ask how far we are going to go in improving Kubernetes scalability. Currently we do not have plans to increase scalability beyond 5000-node clusters (within our SLOs) in the next few releases. If you need clusters larger than 5000 nodes, we recommend to use federation to aggregate multiple Kubernetes clusters.However, that doesn’t mean we are going to stop working on scalability and performance. As we mentioned at the beginning of this post, our top priority is to refine our two existing SLOs and introduce new ones that will cover more parts of the system, e.g. networking. This effort has already started within the Scalability SIG. We have made significant progress on how we would like to define performance SLOs, and this work should be finished in the coming month.Join the effortIf you are interested in scalability and performance, please join our community and help us shape Kubernetes. There are many ways to participate, including:Chat with us in the Kubernetes Slack scalability channel: Join our Special Interest Group, SIG-Scalability, which meets every Thursday at 9:00 AM PSTThanks for the support and contributions! Read more in-depth posts on what’s new in Kubernetes 1.6 here. — Wojciech Tyczynski, Software Engineer, Google[1] We are investigating why 5000-node clusters have better startup time than 2000-node clusters. The current theory is that it is related to running 5000-node experiments using 64-core master and 2000-node experiments using 32-core master.
Quelle: kubernetes

Thimbleweed Park im Test: Mord im Pixelparadies

Nach vielen, vielen Jahren gibt es endlich ein neues Adventure von Lucas Arts – fast jedenfalls: Thimbleweed Park von Ron Gilbert (Monkey Island) sieht aus wie die Klassiker und spielt sich auch so. Das offizielle Firmenlogo fehlt, trotzdem ist der Titel nicht nur für Retrofans ein großer Spaß! (Thimbleweed Park, Spieletest)
Quelle: Golem

The trends fueling hybrid cloud management

What a week at IBM Interconnect 2017. After a busy and productive week at our signature cloud event, I’d like to recap some of the key trends and directions that IBM is seeing in the hybrid cloud management space.
Not long ago, the cloud was seen primarily as a cost-cutting tool. Many saw it as means of improving agility through easy access to low-cost infrastructure. But today, the cloud is gaining traction for its potential to change the dynamics of innovation and digital disruption.
Unlocking more value in the cloud
One way to drive the new experiences clients are after is to bring together value from across the business. This can include a combination of existing data and applications with new methods of engagement in the form of mobile apps, APIs and microservices.
This approach leverages current investments while extending the reach of business right into the hand of their customers. Businesses can also augment internal data with external insights coming from social media, weather forecasts, and Internet of Things (IoT) devices to enrich customer experiences and interactions.
As companies move to more strategic consumption of the cloud, they often uncover opportunities to reimagine business processes and entire industry models. Cognitive services— artificial intelligence—can help businesses learn from individual customer interactions over time. They can adapt automatically to accommodate changing preferences, buying patterns and even understand tone-of-voice to tailor interactions to the person’s mood. On a broader scale, the ability to combine cognitive insights with cloud services can determine the course of industry leaders and laggards.
Setting cloud strategy today for long-term gains
The extent to which business leaders consider the implications of the cloud strategy they select today could have the greatest impact on long-term success. Is the strategy centered solely on scalable infrastructure? Is it adaptable to your business model and investment levels? Does it deliver higher-value business applications and industry functions? Will it enable game-changing business models equipped with blockchain, cognitive services and new data insights to optimize customer experiences?
One key factor in the race to outpace competitors is the ability to innovate with speed. Companies want to select the best of their capabilities and combine it with the latest of what’s available externally from vendors, partners and communities. That’s a critical benefit of using the cloud.
Public clouds enable companies to get new value outside-in from third parties, allowing for rapid adoption of new services, including data, applications, devops toolchains, or community innovations.
Private clouds help companies extract more value inside-out from their business, allowing them to securely extract and analyze data on their customers, transactions and products.
Combining public cloud and private cloud enables business advantage by driving business outcomes more quickly, more effectively and at lower cost. This is why multicloud environments are rapidly becoming the new norm for the enterprise.
Introducing IBM Cloud Automation Manager
The right multicloud strategy can allow companies to combine the delivery and consumption models that best suit their unique business and industry requirements. They can mix and match in a way that optimizes speed, flexibility and business value. To optimize the benefits of these diverse environments requires a unified cloud management platform.
This week, at InterConnect, we were thrilled to unveil Cloud Automation Manager, a purpose-built, cloud-agnostic, multicloud management platform created together with clients to provide significant business value. With Cloud Automation Manager, you can rapidly automate provisioning of cloud applications and resources on any cloud, while leveraging cognitive insights to manage your multicloud environment with ease.
Cloud Automation Manager includes pre-built automation packs spanning infrastructure to full stack apps and helps companies optimize workload placement without lock-in or loss of control. It combines speed, flexibility, control and intelligence so that IT operations managers can more easily and efficiently provision and automate workloads across multiple clouds. At the same time, they can provide developers and DevOps teams with self-service access to a catalog of cloud resources and applications—all from their cloud of choice.
To find out more, please visit our website. To get started with Cloud Automation Manager today at no cost, visit us at ibm.biz/tryIBMCAM.
The post The trends fueling hybrid cloud management appeared first on news.
Quelle: Thoughts on Cloud