Day 1 at Next ‘19: Hybrid cloud, full-stack serverless, open-source partnerships, and more

The first day of Next ‘19 kicked off with a bang—from a visit by Google CEO Sundar Pichai and an in-depth keynote hosted by new Google Cloud CEO Thomas Kurian to hundreds of breakout sessions that gave attendees the tools and knowledge they need to be successful in the cloud.Here are some highlights from day one:Making hybrid cloud a realityWe introduced Anthos—formerly Cloud Services Platform—to help customers by providing a unified programming model and monitoring/policy management framework across on-premises and multiple clouds, not just Google Cloud Platform (GCP). Cisco and VMware joined us on stage to talk about how we’re working together to make hybrid cloud a reality.What we announcedAnthos is now generally available on Google Kubernetes Engine (GKE) and GKE On-Prem, so you can deploy, run and manage your applications on-premises or in the cloud. Coming soon, we’ll extend that flexibility to third-party clouds like AWS and Azure. And Anthos is launching with the support of more than 30 hardware, software and system integration partners so you can get up and running fast.With Anthos Migrate, powered by Velostrata’s migration technology, you can auto-migrate VMs from on-premises or other clouds directly into containers in GKE with minimal effort.Anthos Config Management lets you create multi-cluster policies out of the box that set and enforce role-based access controls, resource quotas, and namespaces—all from a single source of truth.Partners such as VMware, Dell EMC, HPE, Intel, and Lenovo have committed to delivering Anthos on their own hyperconverged infrastructure for their customers. By validating Anthos on their solution stacks, our mutual customers can choose hardware based on their storage, memory, and performance needs.Bringing serverless to the full stackFrom constrained runtime support to vendor lock-in, traditional serverless offerings often come with some significant challenges. As a result, developers often find themselves choosing between the ease and velocity of serverless or the flexibility and portability of containers. Today, we introduced Cloud Run, a new serverless compute platform for containerized apps with portability built-in, to give you the best of both worlds.What we announcedCloud Run, our fully managed serverless execution environment, offers serverless agility for containerized apps.Cloud Run on GKE brings the serverless developer experience and workload portability to your GKE cluster.Knative, the open API and runtime environment, brings a serverless developer experience and workload portability to your existing Kubernetes cluster anywhere.We’re also making new investments in our Cloud Functions and App Engine platforms with new second generation runtimes, a new open-sourced Functions Framework, and additional core capabilities, including connectivity to private GCP resources.Announcing partnerships to bring fully managed open source to Google CloudGoogle Cloud is deeply committed to open source—and we know so many organizations would benefit from fully managed services that let you get the best of open-source innovation and operate those technologies at scale. To help, Thomas announced our strategic partnerships with leaders in the open source community to deliver the industry’s first comprehensive platform for fully managed open source-centric database and analytics services.What we announcedOur strategic partnerships with Confluent, MongoDB, Elastic, Neo4j, Redis Labs, InfluxData, and Datastax tightly integrate their open source-centric technologies into GCP, providing a seamless user experience across management, billing and support.What our customers and partners are sayingWe’re glad to bring our larger community together this week to make connections and learn from each other. We’ve heard some great stories today about how the cloud is helping change the way companies do business, whether they’re improving global patient care or reimagining bank branches as centers for community and education. Here are a few of our favorite quotes from day one:”It’s not about the cloud working backward, it’s about the realities of the environment that all of our customers live in today. We need to build open ecosystems the people around us can embrace and extend.”—David Goeckeler, EVP and General Manager, Networking and Security, Cisco“Retail in the midst of major transformation. As a retailer, we needed to innovate. We also needed to become more engineering focused. Google has that same DNA—an engineering focus driving innovation. That really blended us together in terms of this partnership.” —Ratnakar Lavu, CTO, Kohl’s“Monetizing open source was always a very big challenge for open source vendors and more so in the cloud era. Google has taken a very different approach from other cloud vendors when it comes to open source, and I think this is great news for the open source community.” —Ofer Bengal, CEO, Redis Labs.The start of three great daysWith the day one keynote done, we kicked off more than 500 breakout sessions, cloud talks, and hands-on labs on topics ranging from API management for serverless and multi-cloud and undoing human bias at scale with Kubeflow to a talk on inclusion with actress and equal rights advocate Laverne Cox. Keep an eye on our Google Cloud, GCP, and G Suite YouTube channels for the latest session uploads—we’ll be adding them as fast as we can.Looking forward to a fantastic day two!
Quelle: Google Cloud Platform

How to accelerate DevOps with Machine Learning lifecycle management

DevOps is the union of people, processes, and products to enable the continuous delivery of value to end users. DevOps for machine learning is about bringing the lifecycle management of DevOps to Machine Learning. Utilizing Machine Learning, DevOps can easily manage, monitor, and version models while simplifying workflows and the collaboration process.

Effectively managing the Machine Learning lifecycle is critical for DevOps’ success. And the first piece to machine learning lifecycle management is building your machine learning pipeline(s).

What is a Machine Learning Pipeline? 

DevOps for Machine Learning includes data preparation, experimentation, model training, model management, deployment, and monitoring while also enhancing governance, repeatability, and collaboration throughout the model development process. Pipelines allow for the modularization of phases into discrete steps and provide a mechanism for automating, sharing, and reproducing models and ML assets. They create and manage workflows that stitch together machine learning phases. Essentially, pipelines allow you to optimize your workflow with simplicity, speed, portability, and reusability.

There are four steps involved in deploying machine learning that data scientists, engineers and IT experts collaborate on:

Data Ingestion and Preparation
Model Training and Retraining
Model Evaluation
Deployment

Together, these steps make up the Machine Learning pipeline. Below is an excerpt from documentation on building machine pipelines with Azure Machine Learning service, which explains it well.

“Using distinct steps makes it possible to rerun only the steps you need, as you tweak and test your workflow. A step is a computational unit in the pipeline. As shown in the preceding diagram, the task of preparing data can involve many steps. These include, but aren't limited to, normalization, transformation, validation, and featurization. Data sources and intermediate data are reused across the pipeline, which saves compute time and resources.”

4 benefits of accelerating Machine Learning pipelines for DevOps

 

1. Collaborate easily across teams

Data scientists, data engineers, and IT professionals using machine learning pipelines need to collaborate on every step involved in the machine learning lifecycle: from data prep to deployment.
Azure Machine Learning service workspace is designed to make the pipelines you create visible to the members of your team. You can use Python to create your machine learning pipelines and interact with them in Jupyter notebooks, or in another preferred integrated development environment.

2. Simplify workflows

Data prep and modeling can last days or weeks, taking time and attention away from other business objectives.
The Azure Machine Learning SDK offers imperative constructs for sequencing and parallelizing the steps in your pipelines when no data dependency is present. You can also templatize pipelines for specific scenarios and deploy them to a REST endpoint, so you can schedule batch-scoring or retraining jobs. You only need to rerun the steps you need, as you tweak and test your workflow when you rerun a pipeline.

3. Centralized Management

Tracking models and their version histories is a hurdle many DevOps teams face when building and maintaining their machine learning pipelines.
The Azure Machine Learning service model registry tracks models, their version histories, their lineage and artifacts. Once the model is in production, the Application Insights service collects both application and model telemetry that allows the model to be monitored in production for operational and model correctness. The data captured during inferencing is presented back to the data scientists and this information can be used to determine model performance, data drift, and model decay, as well as the tools to train, manage, and deploy machine learning experiments and web services in one central view.
The Azure Machine Learning SDK also allows you to submit and track individual pipeline runs. You can explicitly name and version your data sources, inputs, and outputs instead of manually tracking data and result paths as you iterate. You can also manage scripts and data separately for increased productivity. For each step in your pipeline. Azure coordinates between the various compute targets you use, so that your intermediate data can be shared with the downstream compute targets easily. You can track the metrics for your pipeline experiments directly in the Azure portal.

4. Track your experiments easily

 

DevOps capabilities for machine learning further improve productivity by enabling experiment tracking and management of models deployed in the cloud and on the edge. All these capabilities can be accessed from any Python environment running anywhere, including data scientists’ workstations. The data scientist can compare runs, and then select the “best” model for the problem statement.
The Azure Machine Learning workspace keeps a list of compute targets that you can use to train your model. It also keeps a history of the training runs, including logs, metrics, output, and a snapshot of your scripts. Create multiple workspaces or common workspaces to be shared by multiple people.

 

Conclusion

As you can see, DevOps for Machine Learning can be streamlined across the ML pipeline with more visibility into training, experiment metrics, and model versions. Azure Machine Learning service, seamlessly integrates with Azure services to provide end-to-end capabilities for the entire Machine Learning lifecycle, making it simpler and faster than ever.

This is part two of a four-part series on the pillars of Azure Machine Learning services. Check out part one if you haven’t already, and be sure to look out for our next blog, where we’ll be talking about ML at scale.

Learn More

Visit our product site to learn more about the Azure Machine Learning service, and get started with a free trial of Azure Machine Learning service.
Quelle: Azure

How do teams work together on an automated machine learning project?

When it comes to executing a machine learning project in an organization, data scientists, project managers, and business leads need to work together to deploy the best models to meet specific business objectives. A central objective of this step is to identify the key business variables that the analysis needs to predict. We refer to these variables as the model targets, and we use the metrics associated with them to determine the success of the project.

In this use case, available to the public on GitHub, we’ll see how a data scientist, project manager, and business lead at a retail grocer can leverage automated machine learning and Azure Machine Learning service to reduce product overstock. Azure Machine Learning service is a cloud service that you use to train, deploy, automate, and manage machine learning models, all at the broad scale that the cloud provides. Automated machine learning within Azure Machine Learning service is the process of taking training data with a defined target feature, and iterating through combinations of algorithms and feature selections to automatically select the best model for your data based on the training scores.

Excess stock quickly becomes a liquidity problem, as it is not converted back to cash unless margins are reduced by means of discounts and promotions or, even worse, when it accumulates to be sent to other channels such as outlets, delaying its sale. Identifying in advance which products will not have the level of rotation they expect and controlling replenishment with stock cover that is aligned with sales forecasts are key factors in helping retailers achieve ROI on their investments. Let’s see how the team goes about solving this problem and how automated machine learning enables the democratization of artificial intelligence across the company.

Identify the right business objective for the company

Strong sales and profits are the result of having the right product mix and level of inventory. Achieving this ideal mix requires having current and accurate inventory information. Manual processes not only take time, causing delays in producing current and accurate inventory information, but also increase the likelihood of errors. These delays and errors are likely to cause lost revenue due to inventory overstocks, understocks, and out-of-stocks.

Overstock inventory can also take valuable warehouse space and tie up cash that ought to be used to purchase new inventory. But selling it in liquidation mode can cause its own set of problems, such as tarnishing your reputation and cannibalizing sales of other current products.

The project manager, being the bridge between data scientists and business operations, reaches out to the business lead to discuss the possibilities of using some of their internal and historical sales to solve their overstock inventory problem. The project manager and the business lead define project goals by asking and refining tangible questions that are relevant for the business objective.

There are two main tasks addressed in this stage:

Define objectives: The project manager and the business lead need to identify the business problems and, most importantly, formulate questions that define the business goals that the data science techniques can target.
Identify data sources: The project manager and data scientist need to find relevant data that helps answer the questions that define the objectives of the project.

Look for the right data and pipeline

It all starts with data. The project manager and the data scientist need to identify data sources that contain known examples of answers to the business problem. They look for the following types of data:

Data that is relevant to the question. Do they have measures of the target and features that are related to the target?
Data that is an accurate measure of their model target and the features of interest.

There are three main tasks that the data scientist needs to address in this stage:

Ingest the data into the target analytics environment
Explore the data to determine if the data quality is adequate to answer the question
Set up a data pipeline to score new or regularly refreshed data

After setting up the process to move the data from the source locations to the target locations where it’s possible to run analytics operations, the data scientist starts working on raw data to produce a clean, high-quality data set whose relationship to the target variables is understood. Before training machine learning models, the data scientist needs to develop a sound understanding of the data and create a data summarization and visualization to audit the quality of the data and provide the information needed to process the data before it's ready for modeling.

Finally, the data scientist is also in charge of developing a solution architecture of the data pipeline that refreshes and scores the data regularly.

Forecast orange juice sales with automated machine learning

The data scientist and project manager decide to use automated machine learning for a few reasons: automated machine learning empowers customers, with or without data science expertise, to identify an end-to-end machine learning pipeline for any problem, achieving higher accuracy while spending far less of their time. And it also enables a significantly larger number of experiments to be run, resulting in faster iteration toward production-ready intelligent experiences.

Let’s look at how their process using automated machine learning for orange juice sales forecasting delivers on these benefits.

After agreeing on the business objective and what type of internal and historical data should be used to meet that objective, the data scientist creates a workspace. This workspace is the top-level resource for the service and provides data scientists with a centralized place to work with all the artifacts they need to create. When a workspace is created in an AzureML service, the following Azure resources are added automatically (if they are regionally available):

Azure Container Registry
Azure Storage
Azure Application Insights
Azure Key Vault

To run automated machine learning, the data scientist also needs to create an Experiment. An Experiment is a named object in a workspace that represents a predictive task, the output of which is a trained model and a set of evaluation metrics for the model.

The data scientist is now ready to load the historical orange juice sales data and loads the CSV file into a plain pandas DataFrame. The time column in the CSV is called WeekStarting, so it will be specially parsed into the datetime type.

Each row in the DataFrame holds a quantity of weekly sales for an orange juice brand at a single store. The data also includes the sales price, a flag indicating if the orange juice brand was advertised in the store that week, and some customer demographic information based on the store location. For historical reasons, the data also includes the logarithm of the sales quantity.

The task is now to build a time series model for the Quantity column. It’s important to note that this data set is comprised of many individual time series; one for each unique combination of Store and Brand. To distinguish the individual time series, we thus define the grain—the columns whose values determine the boundaries between time series.

After splitting the data into a training and a testing set for later forecast evaluation, the data scientist starts working on the modeling step for forecasting tasks, and automated machine learning uses pre-processing and estimation steps that are specific to time series. Automated machine learning will undertake the following pre-processing steps:

Detect the time series sample frequency (e.g., hourly, daily, weekly) and create new records for absent time points to make the series regular. A regular time series has a well-defined frequency and has a value at every sample point in a contiguous time span.
Impute missing values in the target via forward-fill and feature columns using median column values.
Create grain-based features to enable fixed effects across different series.
Create time-based features to assist in learning seasonal patterns.
Encode categorical variables to numeric quantities.

The AutoMLConfig object defines the settings and data for an automated machine learning training job. Below is a summary of automated machine learning configuration parameters that were used for training the orange juice sales forecasting model:

Visit GitHub for more information on forecasting. Each iteration runs within an experiment and stores serialized pipelines from the automated machine learning iterations until they retrieve the pipeline with the best performance on the validation data set.

Once the evaluation has been performed, the data scientist, project manager, and business lead meet again to review the forecasting results. It’s the project manager and business lead’s job to make sense of the outputs and choose practical steps based on those results. The business lead needs to confirm that the best model and pipeline meet the business objective and that the machine learning solution answers the questions with acceptable accuracy to deploy the system to production for use by their internal sales forecasting application.

Microsoft invests in Automated Machine Learning

Automated machine learning is based on a breakthrough from the Microsoft Research division. The approach combines ideas from collaborative filtering and Bayesian optimization to search an enormous space of possible machine learning pipelines intelligently and efficiently. It’s essentially a recommender system for machine learning pipelines. Similar to how streaming services recommend movies for users, automated machine learning recommends machine learning pipelines for data sets.

It’s now offered as part of the Azure Machine Learning service. As you’ve seen here, Automated machine learning empowers customers, with or without data science expertise, to identify an end-to-end machine learning pipeline for any problem and save time while increasing accuracy. It also enables a larger number of experiments to be run and faster iterations. How could automated machine learning benefit your organization? How could your team work more closely on using machine learning to meet your business objectives?

 

Resources

Learn more about Azure Machine Learning service
Learn more about automated machine learning
Get started with a free trial of the Azure Machine Learning service

Quelle: Azure

Building a better cloud together: News from our partners at Next ‘19

Whether they’re looking to modernize their infrastructure, find meaningful insights from their data, or take advantage of industry-specific expertise, our customers rely on partners to help them on their cloud journey. Our Google Cloud ecosystem—including reseller partners, services partners, technology partners, and open source innovators—plays that role, and over the past year we’ve made significant investments in how we support that ecosystem.As we head into Next ’19 this week, we are excited to announce even more updates and investments to continue this momentum into the next year and beyond. These include:Wide-ranging support from our partners for Anthos.A commitment to a robust open source partner community.Expansions of compute and serverless partnerships.New specializations for our reseller and services partners.New partner integrations focused on improving collaboration for end users.Updates to our approach to MSPs.And of course, our 2018 partner award winners.Here’s a round-up of some of our exciting partner news:Making multi-cloud a reality with Anthos and broad support from our ecosystemAnthos, Google Cloud’s new hybrid platform, enables businesses to write their apps once and run them anywhere—in their own data centers, Google Cloud or another cloud.For companies that want to leverage existing investments in their own data centers, we have partnered with Cisco, Dell EMC, HPE, Intel, Lenovo and VMware, each of whom have committed to validating Anthos on their solution stack and providing reference architectures, giving our mutual customers a choice of hardware based on their storage, memory, and performance needs.We also know that enterprises sometimes need help to modernize and extend applications using Anthos. We’ve lined up leading systems integrators who are ready to work with you to incorporate Anthos into your environment, including:But we didn’t stop there. We want to give customers access to an entirely cloud-native ecosystem of open solutions, from managed databases and security to analytics and DevOps tools. Through Anthos, customers can deploy these solutions seamlessly between GKE and GKE On-Prem—in many cases in one-click as Kubernetes applications via GCP Marketplace, which includes integrated metering and billing. These include:You can learn more about Anthos in our blog post.Committing to robust, open-source innovation and partnershipsSince Google’s founding, open source has been at the core of what we do, and you can see this reflected in our contributions to projects like Kubernetes, TensorFlow, Go, Android, and many more. We’ve always seen our friends in the open-source community as equal collaborators, and not simply a resource to be mined.Today, we’re extending our long-standing commitment to, and participation in, the open-source community with a range of strategic partnerships. We’ll be offering managed services operated by open-source centric companies CloudBees, Confluent, MongoDB, Elastic.co, Neo4j, Redis Labs, InfluxData and Datastax that are tightly integrated into Google Cloud Platform (GCP), providing a seamless user experience across management, billing and support. This makes it easier for our enterprise customers to build on open source technologies, and it delivers on our commitment to continually support and grow these open source communities.By offering an elevated experience similar to Google’s native services, we’re working alongside these creators and supporting the growth of these companies’ technologies to inspire adoption. You can learn more in our open source partnership blog post.New serverless and compute partnershipsLast week, we announced the expansion of our Compute Engine virtual machine (VM) offerings to include new Compute-Optimized VMs and Memory-Optimized VMs. Both are based on 2nd Generation Intel Xeon Scalable Processors, which we delivered to customers last October—the first cloud provider to do so. In addition, these processors will also be coming to our general-purpose VMs. This means you’ll have access to a complete portfolio of machine types to successfully run your workloads across a wide range of memory and compute requirements.Google Cloud also became the first Cloud provider to offer availability for the NVIDIA Tesla T4GPU back in November 2018. The T4 GPUs are optimized for machine learning (ML) inference, distributed training of models, and computer graphics.We’re also expanding our ecosystem of serverless partnerships. Serverless computing helps developers focus on writing code that delivers business value, rather than worrying about the complicated underlying infrastructure. Our unique approach to serverless makes it easy for you to tap into a full stack of additional services and an entire ecosystem of deep partnerships with industry leaders such as Datadog, NodeSource, GitLab, and StackBlitz, which provide integration support for Cloud Run across application monitoring, coding, and deployment stages. To learn more about our new serverless offerings, check out “Announcing Cloud Run, the newest member of our serverless compute stack”.Expanded ways to collaborateSharing and storing web-based files has quickly become an indispensable part of day-to-day work, but corralling those files across multiple tools can be difficult. We want to make it easy for you to collaborate no matter the file, no matter the app. Today, we announced a new integration with G Suite and Dropbox to make this easier.Now, Dropbox Business customers can create, save and share G Suite files—like Google Docs, Sheets and Slides—right from Dropbox. With this integration, G Suite files will show up when you search in Dropbox. You can also open up different file formats like .docx, .xlsx or .pptx files from Dropbox in G Suite and then save them back to Dropbox in whichever format you prefer. (While we’re on the subject of file formats, we’ve also made it easier to collaborate in Microsoft Office files in real-time. Read more about that in this post).Expanding our partner specializationsMany customers need partners with experience in a particular area to help them build advanced solutions to their challenges, and run and maintain them over time.To help customers identify the right partner, we provide partner specializations. We offer 10 different partner specialization areas, a significant range that allows customers to search for the right blend of relevant depth to address their specific needs and requirements. Partners who have earned a specialization have multiple individuals on staff with the highest level of certification, demonstrated and documented customer success stories, and have passed a rigorous capability assessment from experts on our professional services team.Today we are announcing that 21 of our partners have achieved specializations in three additional areas. These include:Marketing Analytics Specialization Partners, with the unique skills to collect, transform, analyze and visualize data, and then use the insights gained to optimize marketing strategy and activations.55 SASAccentureBrainPadDatatonicDeloitteMaven Wave PartnersMerkleNRIQuantiphi, IncPublicis.SapientServianWunderman ThompsonIoT Specialization Partners, which help businesses more easily and securely connect, manage, and ingest IoT data from globally dispersed devices, in order to analyze it and gain new insights. Partners announced today include:Agosto, IncAtos ITHitachi ConsultingLeverege, LLCObject Computing Inc.Softserve Inc.Security Training Specialization Partners, who have consistently delivered successful training according to Google Cloud curriculum standards, along with a minimum of two authorized trainers dedicated to the security track of GCP. These include:ROI TrainingQA LimitedJellyfish Group LimitedYou can learn more about our specialization program for partners here.Momentum and updates with our MSPsOur ecosystem of MSPs are a key part of our go-to-market approach. Today, they are helping many customers build, scale, manage and support modernized IT strategies with Google Cloud at their core.Qualified Google Cloud MSPs have completed trainings and extensive verifications to ensure they have the ability to manage, monitor, remediate and optimize workloads on GCP. We provide these partners with numerous resources to help them serve customers, including support from dedicated MSP technical account managers, customer engineering, and online and in-person community resources.The response from the MSP community and from customers has been strong. Over the last year, we have qualified partners including Rackspace, DoIT, Infosys, and many more. At Next, we will be introducing an MSP Initiative badge for qualified partners, making it easier for our joint customers to discover partners who can help them to accelerate their Google Cloud journey.To learn more about our MSP initiative, please visit here.Announcing the winners of our 2018 Partner AwardsEach year we look forward to the opportunity to recognize our amazing partners helping our customers at every stage of their journey to the cloud. At 5pm PT today, during our Partner Summit keynote, we’ll be announcing our 2018 winners and we’ll be updating this post with the full list—so keep an eye out.
Quelle: Google Cloud Platform

Snapdragon-Chips: Qualcomm kündigt 665 und 730(G) an

Mit dem Snapdragon 665, dem Snapdragon 730 und dem Snapdragon 730G hat Qualcomm weitere Smartphone-Chips vorgestellt. Der 665 ist schneller bei der Berechnung künstlicher Intelligenz, der 730 weist viele Funktionen des 855 auf und der 730G wurde für Spiele optimiert. (Snapdragon, Smartphone)
Quelle: Golem