Managing the Looker ecosystem at scale with SRE and DevOps practices

Many organizations struggle to create data-driven cultures where each employee is empowered to make decisions based on data. This is especially true for enterprises with a variety of systems and tools in use across different teams. If you are a leader, manager, or executive focused on how your team can leverage Google’s SRE practices or wider DevOps practices, definitely you are in the right place!What do today’s enterprises or mature start-ups look like?Today large organizations are often segmented into hundreds of small teams which are often working around data in the magnitude of several petabytes and in a wide variety of raw forms. ‘Working around data’ could mean any of the following: generating, facilitating, consuming, processing, visualizing or feeding back into the system. Due to a wide variety of responsibilities, the skill sets also vary to a large extent. Numerous people and teams work with data, with jobs that span the entire data ecosystem:Centralizing data from raw sources and systemsMaintaining and transforming data in a warehouseManaging access controls and permissions for the dataModeling dataDoing ad-hoc data analysis and explorationBuilding visualizations and reportsNevertheless, a common goal across all these teams is keeping services running and downstream customers happy. In other words, the organization might be divided internally, however, they all have the mission to leverage the data to make better business decisions. Hence, despite silos and different subgoals, destiny for all these teams is intertwined for the organization to thrive. To support such a diverse set of data sources and the teams supporting them, Looker supports over 60 dialects (input from a data source) and over 35 destinations (output to a new data source).Below is a simplified* picture of how the Looker ecosystem is central to a data-rich organization.Simplified* Looker ecosystem in a data-rich environment*The picture hides the complexity of team(s) accountable for each data source. It also hides how a data source may have dependencies on other sources. Looker Marketplace can also play an important role in your ecosystem.What role can DevOps and SRE practices play?In the most ideal state, all these teams will be in harmony as a single-threaded organization with all the internal processes so smooth that everyone is empowered to experiment (i.e. fail, learn, iterate and repeat all the time). With increasing organizational complexities, it is incredibly challenging to achieve such a state because there will be overhead and misaligned priorities. This is where we look up to the guiding principles of DevOps and SRE practices. In case you are not familiar with Google SRE practices, here is a starting point. The core of DevOps and SRE practices are mature communication and collaboration practices. Let’s focus on the best practices which could help us with our Looker ecosystem.Have joint goals. There should be some goals which are a shared responsibility across two or more teams. This helps establish a culture of psychological safety and transparency across teams.Visualize how the data flows across the organization. This enables an understanding how each team plays their role and how to work with them better.Agree on theGolden Signals (aka core metrics). These could mean data freshness, data accuracy, latency on centralized dashboards etc. These signals allow teams to set their error budgets and SLIs.Agree on communication and collaboration methods that work across teams. Regular bidirectional communication modes – have shared Google Chat spaces/slack channels. Focus on artifacts such as jointly owned documentations pages, shared roadmap items, reusable tooling, etc. For example, System Activity Dashboards could be made available to all the relevant stakeholders and supplemented with notes tailored to your organization.Set up regular forums where commonly discussed agenda items include major changes, expected downtime and postmortems around the core metrics. Among other agenda items, you could define/refine a common set of standards, for example centrally defined labels, group_labels, descriptions, etc. in the LookML to ensure there is a single terminology across the board.Promote informal sharing opportunities such as lessons learned, TGIFs, Brown bag sessions, and shadowing opportunities. Learning and teaching have an immense impact on how teams evolve. Teams often become closer with side projects that are slightly outside of their usual day-to-day duties.Have mutually agreed upon change management practices. Each team has dependencies so making changes may have an impact on other teams. Why not plan those changes systematically? For example, getting common standards across the Advance deploy mode.Promote continuous improvements. Keep looking for better, faster, cost-optimized versions of something important to the teams.Revisit your data flow. After every major reorganization, ensure that organizational change has not broken the established mechanisms.despite silos and different subgoals, destiny for all these teams is intertwined for the organization to thrive.Are you over-engineering?There is a possibility that in the process of maturing the ecosystem, we may end up in an overly engineered system – we may unintentionally add toil to the environment. These are examples of toil that often stem from communication gaps. Meetings with no outcomes/action plans – This one is among the most common forms of toil, where the original intention of a meeting is no longer valid but the forum has not taken efforts to revisit their decision.Unnecessary approvals – Being a single threaded team can often create unnecessary dependencies and your teams may lose the ability to make changes.Unaligned maintenance windows – Changes across multiple teams may not be mutually exclusive hence if there is misalignment then it may create unforeseen impacts on the end user.Fancy, but unnecessary tooling – Side projects, if not governed, may create unnecessary tooling which is not being used by the business. Collaborations are great when they solve real business problems, hence it is also required to refocus if the priorities are set right.Gray areas – When you have a shared responsibility model, you also may end up in gray areas which are often gaps with no owner. This can lead to increased complexity in the long run. For example, having the flexibility to schedule content delivery still requires collaboration to reduce jobs with failures because it can impact the performance of your Looker instance.Contradicting metrics – You may want to pay special attention to how teams are rewarded for internal metrics. For example, if a team focuses on accuracy of data and other one on freshness then at scale they may not align with one another.ConclusionTo summarize, we learned how data is handled in large organizations with Looker at its heart unifying a universal semantic model. To handle large amounts of diverse data, teams need to start with aligned goals and commit to strong collaboration. We also learned how DevOps and SRE practices can guide us navigate through these complexities. Lastly, we looked at some side effects of excessively structured systems. To go forward from here, it is highly recommended to start with an analysis of how data flows under your scope and how mature the collaboration is across multiple teams.Further reading and resourcesGetting to know Looker – common use casesEnterprise DevOps GuidebookKnow thy enemy: how to prioritize and communicate risks—CRE life lessonsHow to get started with site reliability engineering (SRE)Bring governance and trust to everyone with Looker’s universal semantic modelRelated articlesHow SREs analyze risks to evaluate SLOs | Google Cloud BlogBest Practice: Create a Positive Experience for Looker UsersBest Practice: LookML Dos and Don’ts
Quelle: Google Cloud Platform

Top 5 Takeaways from Google Cloud’s Data Engineer Spotlight

In the past decade, we have experienced an unprecedented growth in the volume of data that can be captured, recorded and stored.  In addition, the data comes in all shapes and forms, speeds and sources. This makes data accessibility, data accuracy, data compatibility, and data quality more complex than ever more. Which is why this year at our Data Engineer Spotlight, we wanted to bring together the Data Engineer Community to share important learning sessions and the newest innovations in Google Cloud. Did you miss out on the live sessions? Not to worry – all the content is available on demand. Interested in running a proof of concept using your own data? Sign up here forhands-on workshop opportunities.Here are the five biggest areas to catch up on from Data Engineer Spotlight, with the first four takeaways written by a loyal member of our data community: Francisco Garcia, Founder of Direcly, a Google Cloud Partner. #1: The next generation of Dataflow was announced, including Dataflow Go (allowing engineers to write core Beam pipelines in Go, data scientists to contribute with Python transforms, and data engineers to import standard Java I/O connectors). The best part, it all works together in a single pipeline. Dataflow ML (deploy easy ML models with PyTorch, TensorFlow, or stickit-learn to an application in real time), and Dataflow Prime (removes the complexities of sizing and tuning so you don’t have to worry about machine types, enabling developers to be more productive). Read on the Google Cloud Blog: The next generation of Dataflow: Dataflow Prime, Dataflow Go, and Dataflow MLWatch on Google Cloud YouTube: Build unified batch and streaming pipelines on popular ML frameworks #2: Dataform Preview was announced (Q3 2022), which helps build and operationalize scalable SQL pipelines in BigQuery. My personal favorite part is that it follows software engineering best practices (version control, testing, and documentation) when managing SQL. Also, no other skills beyond SQL are required. Dataform is now in private preview. Join the waitlist Watch on Google Cloud YouTube: Manage complex SQL workflows in BigQuery using Dataform CLI #3: Data Catalog is now part of Dataplex, centralizing security and unifying data governance across distributed data for intelligent data management, which can help governance at scale. Another great feature is that it has built-in AI-driven intelligence with data classification, quality, lineage, and lifecycle management.  Read on the Google Cloud Blog: Streamline data management and governance with the unification of Data Catalog and Dataplex Watch on Google Cloud YouTube: Manage and govern distributed data with Dataplex#4: A how-to on BigQuery Migration Services was covered, which offers end-to-end migrations to BigQuery, simplifying the process of moving data into the cloud and providing tools to help with key decisions. Organizations are now able to break down their data silos. One great feature is the ability to accelerate migrations with intelligent automated SQL translations.  Read More on the Google Cloud Blog: How to migrate an on-premises data warehouse to BigQuery on Google Cloud Watch on Google Cloud YouTube: Data Warehouse migrations to BigQuery made easy with BigQuery Migration Service #5: The Google Cloud Hero Game was a gamified three hour Google Cloud training experience using hands-on labs to gain skills through interactive learning in a fun and educational environment. During the Data Engineer Spotlight, 50+ participants joined a live Google Meet call to play the Cloud Hero BigQuery Skills game, with the top 10 winners earning a copy of Visualizing Google Cloud by Priyanka Vergadia. If you missed the Cloud Hero game but still want to accelerate your Data Engineer career, get started toward becoming a Google Cloud certified Data Engineer with 30-days of free learning on Google Cloud Skills Boost. What was your biggest learning/takeaway from playing this Cloud Hero game?It was brilliantly organized by the Cloud Analytics team at Google. The game day started off with the introduction and then from there we were introduced to the skills game. It takes a lot more than hands on to understand the concepts of BigQuery/SQL engine and I understood a lot more by doing labs multiple times. Top 10 winners receiving the Visualizing Google Cloud book was a bonus. – Shirish KamathCopy and pasting snippets of codes wins you competition. Just kidding. My biggest takeaway is that I get to explore capabilities of BigQuery that I may have not thought about before. – Ivan YudhiWould you recommend this game to your friends? If so, who would you recommend it to and why would you recommend it? Definitely, there is so much need for learning and awareness of such events and games around the world, as the need for Data Analysis through the cloud is increasing. A lot of my friends want to upskill themselves and these kinds of games can bring a lot of new opportunities for them. – Karan KukrejaWhat was your favorite part about the Cloud Hero BigQuery Skills game? How did winning the Cloud Hero BigQuery Skills game make you feel?The favorite part was working on BigQuery Labs enthusiastically to reach the expected results and meet the goals. Each lab of the game has different tasks and learning, so each next lab was giving me confidence for the next challenge. To finish at the top of the leaderboard in this game makes me feel very fortunate. It was like one of the biggest milestones I have achieved in 2022. – Sneha Kukreja
Quelle: Google Cloud Platform

Get to know the top 3 teams of the Google Cloud Hackathon Singapore

Google Cloud hackathonOn 10th April 2022, Google Cloud launched the first Singapore Google Cloud Hackathon, where startup teams were tasked to build solutions for either the topics of Sustainability, Artificial Intelligence, Automation or the New Normal, to create innovative solutions and have the opportunity to win prizes. From April to 10th June, Google Cloud worked with hackathon entrants through the solutioning process from ideation to prototyping to the final pitch. The hackathon saw incredible response with 40 startup teams competing for a top 5 spot. The top 5 teams were invited to pitch live at the Google Asia Pacific Singapore Campus and presented to a panel of judges that consisted of experts in the startup ecosystem and technology leaders across APAC in Google. The top 3 teams also continued to receive mentorship opportunities with Google Cloud and startup experts.  Top 3 teamsRead on to learn more about the top 3 startup teams:Team Empathly – 2nd Runner UpCofounders Timothy Liau, Jamie Yau and Rachel Tan personally experienced hate speech and witnessed discrimination in online communities. The available content filters and manual moderation solutions, which they found to be very expensive, only focused on damage control after the hateful comment has been sent. Out of a desire to prevent the toxic behavior at its source, Empathly was born.Any platform with user-generated content — social media, games, marketplaces, dating apps and more — is susceptible to hate speech. Described as “The Grammarly for content moderation”, Empathly applies its AI that identifies distinct types of hate speech with context – to promote safer and more inclusive speech in workplaces and online platforms. Empathy is built on Cloud Run and Cloud Firestore.Empathly’s behavioral science advisory team includes Yale-NUS professor and expert in behavioral insights Dr. Jean Liu whose research focuses on how technological solutions require an appreciation of human behavior and the social context. They will focus their next few months on working closely with their early customers and building toward product-market-fit.Team Ambient Systems – 1st Runner UpIvan Damnjanović founded Ambient to help companies meet their decarbonization targets through data science innovation. The team consists of Ivan and Frey Liu, a fellow computer science masters student from National University of Singapore (NUS). Through Ambient’s platform, companies can access real time Big Data analytics for actionable decarbonisation through energy efficiency and trade-off optimization.In 2020, Ambient Systems was founded when Ivan proposed a software-based solution for managing complex indoor air quality challenges, such as airborne transmission of COVID and vertical farming climate conditions. Ivan’s patent in the AgriTech field helped Ambient secure a $100k investment from NUS to further pursue commercialization of the technology and help Singapore achieve its 30 by 30 agenda – to build Singapore’s “agri-food industry’s capability and capacity to produce 30% of our nutritional needs locally and sustainably by 2030”.Through the use of Google’s Firebase platform, the team was able to quickly build a fully functional prototype that garnered interest from investors and customers.Team Pomona – ChampionsAs harsh weather conditions continue to plummet agricultural yield, there is an increasing need for countries to improve food security through efficient agriculture and sustainable living. However, high operational costs and cyclical risks inhibit the growth of vertical farming in the agricultural industry.Team Pomona consists of Pang Jun Rong, Yuen Kah May, Teo Keng Swee, Nicole Lim Jia Yi, Tan Jie En, who are student entrepreneurs from Singapore Management University (SMU) Computer Science. They took motivation from their school’s efforts in sustainability and technology to form Pomona — a solution set on making food security more personal through the ownership of vegetables in commercial agricultural lifecycles.Pomona features a gamified de-fi agricultural platform to promote collective ownership of vertical farming agriculture, which enables profit-sharing between producers and consumers to hedge against operational risks. This was done through a hybrid decentralized microservice cloud architecture with Google Cloud, using blockchain technologies for “dVeg” digital tokens and conventional full-stack components for gamification with IoT integration, providing real-time interactive growth tracking for lifecycle traceability.Final wordsCongratulations to all of the teams and especially Empathy, Ambient Systems, and Pomona. We look forward to more events with startups in the future!
Quelle: Google Cloud Platform

Docker Captain Take 5 — Julien Maitrehenry

Docker Captain Take 5 — Julien Maitrehenry

Docker Captains are select members of the community that are both experts in their field and are passionate about sharing their Docker knowledge with others. “Docker Captains Take 5” is a regular blog series where we get a closer look at our Captains and ask them the same broad set of questions ranging from what their best Docker tip is to whether they prefer cats or dogs (personally, we like whales and turtles over here). Today, we’re interviewing Julien who recently joined the Captains Program. He is a developer and devops at Kumojin and is based in Quebec City.

How/when did you first discover Docker?
I don’t remember how, but, back in 2014, I was working in a PHP shop, and we had our dev environment running inside a VM (with Vagrant). But, it wasn’t really easy to share and maintain. So, we started experimenting with Docker for our dev env. After that, I learned about (legacy/classic) Swarm in Docker, and these tools resolved some of my issues in production for handling load balancer reconfiguration, deployment, and new version management. Check out my first conference about Docker here.
Since then, I continue to learn, use, and share about Docker. It’s useful in so many use cases — it’s amazing!
What is your favorite Docker command?
Docker help. I still need to check the documentation sometimes!
But also, when I need more space on Docker Desktop, I’ll choose docker buildx prune!
What is your top tip for working with Docker that others may not know?
If your team uses ARM (Apple M1, for example) and intel CPU and you want to share a docker image with your team, build a cross platform image:
docker buildx build –push –platform linux/arm64/v8,linux/amd64 –tag xxx/nginx-proxy:1.21-alpine
What’s the coolest Docker demo you have done/seen?
Back during the 2018 Dockercon, I was impressed by the usage of Docker for the DART (Double Asteroid Redirection Test) project. A project as easy as building an aircraft, hitting an asteroid with it, and saving the world!
You should check how they use Docker for space hardware emulation and testing — it’s brilliant to see how Docker could be used to help save the world: https://www.youtube.com/watch?v=RnWXOAplvjY
What have you worked on in the past six months that you’re particularly proud of?
Being a mentor for a developer school in Quebec (42 Quebec). It’s amazing to see the new generation of developers and help them with all their questions, fears, and concerns! And it’s cool when someone calls you “Mister Docker” because he watches a docker conference I gave to answer questions about usage and more.
What do you anticipate will be Docker’s biggest announcement this year?
After Docker Extension and SBOM? It’s really hard to say. I need more time to explore and create my first extension, but, I’m sure the Docker team will find something.
What are some personal goals for the next year with respect to the Docker community?
Give my first conference in English as I always give them in French. I’d also like to update my blog with more content.
What was your favorite thing about DockerCon 2022?
The French community room. It was a pleasure to engage with Aurélie and Rachid and have so many great speakers with us! I would do it again anytime!
Looking to the distant future, what is the technology that you’re most excited about and that you think holds a lot of promise?
7 years from now, I still think Docker will continue to innovate and find new ways to simplify the life of the developer community!
Rapid fire questions…
What new skill have you mastered during the pandemic?
Using a face mask and forgetting about it! Or traveling between different countries during the pandemic with all different kinds of restrictions and rules.
Cats or Dogs?
Cats! I’m sorry, but a dog requires too much time, and I already have 3 young kids.
Salty, sour or sweet?
Salty or Umami
Beach or mountains?
Mountains!
Your most often used emoji?
🤣 or 😄
Quelle: https://blog.docker.com/feed/

Amazon RDS Proxy unterstützt jetzt Amazon RDS für Maria DB in der Version 10.3, 10.4 oder 10.5

Amazon RDS Proxy, ein vollständig verwalteter, hochgradig verfügbarer Datenbankproxy für Amazon Relational Database Service (RDS), unterstützt jetzt Datenbanken von Amazon RDS für MariaDB, die in den Hauptversionen 10.3, 10.4 oder 10.5 ausgeführt werden. Mit Amazon RDS Proxy können Kunden Anwendungen skalierbarer, widerstandsfähiger gegen Datenbankausfälle und sicherer machen.
Quelle: aws.amazon.com

Bekanntgabe der Wiedereinführung von AWS-Sicherheitskompetenz

Wir freuen uns, bekannt zu geben, dass AWS-Sicherheitskompetenz mit neuen konsolidierten Kategorien wieder eingeführt wurde, damit Kunden von AWS validierte AWS-Partnerlösungen leichter finden können. Partner mit Sicherheitskompetenz bieten Lösungen, mit denen Kunden in jeder Phase auf Ihrem Weg in die Cloud die Sicherheit in der Cloud erhöhen können. Von den acht neuen konsolidierten Kategorien enthalten sechs eine kostenlose Sammlung von AWS-validierten Partner-Software- und -Serviceangeboten, die Kunden bei ihren Cloud-Sicherheitssoftware-Toolentscheidungen und mit erweiterten Implementierungs- und Trainingsservices unterstützen: Anwendungssicherheit, Compliance und Privatsphäre, Datenschutz, Identitäts- und Zugriffsmanagement, Infrastrukturschutz, Bedrohungserkennung und -reaktion.
Quelle: aws.amazon.com

Informationen zum Hauptansprechpartner können jetzt in AWS-Konten programmatisch verwaltet werden

Heute machen wir es Kunden einfacher, die Informationen zum Hauptansprechpartner in ihren AWS-Konten mit der AWS Command Line Interface (CLI) und dem AWS SDK anzuzeigen und zu aktualisieren. Wir haben zuvor das Accounts SDK veröffentlicht, mit dem Kunden Fakturierung, Operationen und Sicherheitskontakte für ihre Konten programmatisch verwalten können. Ab heute können Kunden dasselbe SDK verwenden, um auch die Informationen zum Hauptansprechpartner zu aktualisieren, was mit weniger Zeitaufwand verbunden ist als in der Managementkonsole.
Quelle: aws.amazon.com

AWS Fault Injection Simulator unterstützt jetzt ChaosMesh- und Litmus-Experimente

AWS Fault Injection Simulator (FIS) unterstützt jetzt ChaosMesh- und Litmus-Experimente für containerisierte Anwendungen, die in Amazon Elastic Kubernetes Service (EKS) ausgeführt werden. Mit der neuen benutzerdefinierten Kubernetes-Ressourcenaktion für AWS FIS können Sie ChaosMesh- und Litmus-Chaosexperimente innerhalb eines AWS-FIS-Experiments kontrollieren. Dadurch können Sie Fehlereinstreuungs-Workflows in mehreren Tools koordinieren. Sie können beispielsweise einen Stresstest auf der CPU eines Pods mit ChaosMesh- oder Litmus-Fehlern ausführen, während Sie einen zufällig ausgewählten Prozenzsatz an Cluster-Knoten mit AWS-FIS-Fehleraktionen beenden.
Quelle: aws.amazon.com

Der AI Use Case Explorer ist jetzt verfügbar

Der AI Use Case Explorer ist ein geschäftszielorientiertes Web-Suchtool, mit dem Benutzer leicht die richtigen Anwendungsfälle für künstliche Intelligenz (KI) finden, relevante Kundenerfolgsgeschichten entdecken und ihre Teams für KI-Bereitstellungen mobilisieren können. Das benutzerfreundliche Tool nimmt Beschreibungen von Geschäftsproblemen als Eingabe entgegen und stellt relevante, praktische Anwendungsfälle und Geschichten als Ausgaben bereit.
Quelle: aws.amazon.com