Ladies and gentlemen, start your logging

This post was authored by Matías Quaranta, Azure MVP, Autocosmos.

In this blog post, I’m going to show you how I migrated from ELK to Azure Log Analytics and lowered our operation costs by more than ninety percent and reduced our maintenance time.

Background

The need for logging is probably as old as computers and its importance has grown hand in hand with the complexity of distributed architectures.

It is common nowadays for applications and platforms to span multiple servers, service instances, languages and even technologies. Keeping the status and logs of every part becomes a challenge.

In Autocosmos we work entirely on Azure, our whole architecture runs on a myriad of different Azure services, ranging from the most common ones like Azure App Services, Azure Redis Cache and Azure SQL Database to Azure Search, Azure DocumentDB and Microsoft Cognitive Services. We are a technology team that focuses on creating and deploying the best products and platforms for the Automotive and Car enthusiasts in Latin America by leveraging Azure’s SaaS and PaaS offerings.

A few years ago when we decided to implement a centralized logging platform, there weren’t a lot of options that could manage our diverse log output and structure, so we ended up implementing an ELK (Elasticsearch-Logstash-Kibana) Stack running on Azure Virtual Machines. I have to admit, it worked quite well and we were able to absorb JSON logs through Logstash and visualize through Kibana.

Pain points

Like I mentioned, we are a focused technology team of developers and engineers, and even though the ELK Stack worked, we were forced to maintain the virtual Linux environments, watch for Elasticsearch cluster health and a lot of tasks that we, as developers, really didn’t care for. We felt it was becoming a time sink to maintain our logging architecture and it wasn’t truly adding value to our products.

Adding to that, Azure VMs are an IaaS offering, which makes any kind of scaling and network balancing a tedious and complex task for a developer. And all this effort just for our logging architecture.
Who would have thought that a videogame would open the door for a solution to our problems?

Unexpected opportunities

After reading how Halo 5 managed its logging pipeline, we got in touch with the amazing team behind Azure Log Analytics and shared with them our current scenario.

For those not familiar with Azure Log Analytics, it’s a service part of Microsoft Operations Management Suite but has a separate pricing (including a free tier) and allows for collection, storing and analysis of log data from multiple sources, which includes Windows and Linux environment sources (on-premises or cloud).

To our surprise, they were already working on an implementation that would allow for custom log ingestion that wasn’t exclusively coming from a declared agent or source, through an HTTP API and using JSON objects.

This was exactly the solution we needed! We were already using JSON for our ELK Stack, so it was as easy as redirecting our log flow to Azure Log Analytics HTTP Data Collector API. Our whole architecture started directing its logs, taking advantage of the freedom of HTTP and JSON and the fact that Azure Log Analytics parses and understands each JSON attribute separately, we could just send our Front End logs, our DocumentDB logs, Cognitive Services logs. We could even send logs from inside Azure Functions, though they were different sources with different information, it just worked!

It became the backbone of all our operational insights.

Implementing JSON logging by HTTP

Every service and application that is part of our solution is running ASP.NET Core, so we created a .Net wrapper for the Azure Log Analytics HTTP Data Collector API in the form of a Nuget Package open to contributions on a GitHub public repository. This wrapper lets us send JSON payloads from any part of our architecture and even supports object serialization.

We use logging for post-mortem analysis, achievable by directly sending the Exception when an internal call on our APIs or Webjob function fails and then using the search experience in Azure Log Analytics to find the root of the problem:

Performance can be also measured with logs; we track our DocumentDB Request Units used on each call (using the exposed headers) to measure how our hourly quota is used and alert us if we are reaching a point where we need to scale:

Search experience

Azure Log Analytics is not just a log storage, it includes a powerful Search feature that lets you delve deep into your log data with a well-documented query syntax and shows each JSON attribute parsed and filterable:

 

You can even create custom Dashboards with widgets and tiles of your choice and build the visualizations best-suited for your scenario:

Finally, you can configure Alerts based on Searches and time frames and tied them with Webhooks, Azure Automation Runbooks or Email notifications.

Conclusion

Implementing Azure Log Analytics meant that not only we reduced our maintenance time, we also lowered our operation costs (no more VMs!) by more than ninety percent. All our time now devoted to what matters most for our business, creating and building the best products and platforms for Car enthusiasts in Latin America.
Quelle: Azure

Microsoft Ignite: Learn about Azure Stack infrastructure operations and management

We are thrilled to host you at Ignite next week. For us on the engineering team, it&;s an opportunity to spend time, share and learn from you. This year we have a lot to share, particularly about Azure Stack. We have tried our best to ensure the sessions are structured in a logical progression. Wale authored a great blog post that walked through the outline. The chart below, summarizes how our sessions will be organized.

I am honored to be providing a first look at what we are doing to enable you to manage and operate Azure Stack in your datacenter in the session titled BRK2188: Learn about Azure Stack infrastructure operations and management. With Azure Stack, our goal is to bring Azure into your datacenters, under your operational control. Azure itself is operated by Microsoft and we have a dedicated team of people who use a variety of operational tools and processes for its upkeep. With Azure Stack, we have to provide a facsimile of such tools so you can own and operate it. That is the gist of what we will cover in this session. This starts from how we are working with our partners (Dell, HPE and Lenovo) to deliver integrated systems and provides an overview of the operational tooling that we will provide as a part of Azure Stack.

We shared our thoughts last month in a short blog post and video. I am sure that it has whet your appetite to dig deeper and this is your chance. Start your journey in understanding how we have conceived of Azure Stack being deployed, integrated, monitored, patched/updated and in general how you would manage its lifecycle. After attending this session, you should attend the architecture session, then head to session that will start your journey to turn you into an Azure Stack Infrastructure rock star. A 9:00 am session on the day after the party is always a fun one, but I know your passion and I can be pretty loud, so we will make it work. Look forward to seeing you all there.

Two interesting tidbits, first watch out for this T-Shirt at Ignite. Anyone wearing this, is a part of the Azure Stack team, so stop by and say hello. And second, those blinky lights are saying “I am almost running TP2.”

 

Quelle: Azure

Cognitive computing is driving ROI on cloud

A driving force in cloud adoption is the cost savings one gets from eliminating physical servers, then migrating applications and data into the cloud. This reduces up-front capital expenditures (CAPEX) and enables budget planning that is based around operating expenditures (OPEX) and on-demand service provisioning.
Undoubtedly this propels pricing competition among major cloud service providers. They are striving to reduce their price per virtual machine (VM), which averages 10 percent a year or roughly 5 cents an hour per virtual instance. The continuous pricing reduction derived from multi-tenant public cloud has contributed to the exponential growth of the cloud market for the past five years.
However, from an enterprise user perspective, is the velocity of cost savings enough to be highly competitive in the long term? The convergence of mobile, social, Internet of Things (IoT), cognitive computing and cloud has already disrupted industries by creating innovative solutions around smart cars, smart buildings, smart health care and smart education.
App-based ride sharing companies are now multi-billion-dollar concerns. How can existing taxi service providers around the world sustain their business? Obviously, cost savings from cloud migration is not enough, but they may be compelled to create innovative solutions using cognitive capabilities on cloud.
“Cognitive on cloud” refers to cognitive services running in the cloud and that are available to be consumed via representational state transfer (REST) APIs. These services are available as part of platform-as-a-service (PaaS) offerings such as Bluemix so they can be easily bound to an application while coding.
For example, cognitive analytics such as voice (tone analyzer, speech-to-text) and video (face detection, visual recognition) capabilities enable users to analyze petabytes of unstructured data generated by mobile devices every day.
Developing cognitive applications to run on mobile devices will provide new insights and help organizations create totally new revenue streams.  The convergence of cognitive computing and cloud is driving this cognitive-oriented digital economy. The potential return is seemingly unlimited.

From an return-on-investment (ROI) perspective, let’s suppose that there are two organizations, each planning its cloud adoption. One picks the lowest-priced cloud service provider based on cloud commodity services as well as total cost of ownership (TCO) in the cloud (this is Org 1 in the chart). The other picks a cloud service provider based on one more evaluation criterion in addition to the TCO comparison: cognitive capability (Org 2 in the chart).
As shown in the chart above, by leveraging cognitive capabilities in the cloud, Org 2 will achieve a higher ROI through continuous innovation. The difference between the two organizations is that Org 2 sees cloud as an innovation platform for high ROI as rather than a way to cost savings based on a TCO comparison.
Continuous innovation will also include added values, such as IoT and blockchain powered by the cloud. However, the return may not be immediately quantifiable. Therefore, the gap between the two organizations can be enormous depending on how effectively an organization creates innovative solutions via cognitive applications and other value-added services.
When it comes to selecting the right service provider, ROI on cloud requires more than just a TCO comparison based on the number of VMs, storage capacity, hypervisor, operating system and so on. In addition to this basic analysis, an organization must consider which cloud is cognitive enabled or disabled at the PaaS layer. Achieving a high ROI requires a cognitive-enabled cloud as a foundation. More than 30 IBM Watson services on cloud with a supporting foundationhave become key to the solution for organizations across many industries.
Learn more about cognitive cloud and Watson services at IBM World of Watson.  
The post Cognitive computing is driving ROI on cloud appeared first on news.
Quelle: Thoughts on Cloud