AWS Marketplace für Container unterstützt jetzt die direkte Bereitstellung auf EKS-Clustern

Amazon-EKS-Kunden können jetzt Betriebssoftware von Drittanbietern über die EKS-Konsole oder mithilfe von CLI, eksctl, AWS-APIs oder Tools für Infrastructure as Code wie AWS CloudFormation und Terraform finden und auf ihren EKS-Clustern bereitstellen. Kunden können zwischen kommerzieller, kostenloser oder Open-Source-Software wählen, die Anwendungsfälle wie Überwachung, Sicherheit und Speicherung abdeckt, und dieselben einfachen Befehle verwenden, die sie heute für die Bereitstellung von EKS-Add-ons verwenden, um diese Software von Drittanbietern einzusetzen. Dies hilft EKS-Kunden, den Zeitaufwand für die Suche, das Abonnement und die Bereitstellung von Software von Drittanbietern zu reduzieren, und hilft ihnen, produktionsbereite EKS-Cluster in wenigen Minuten einzurichten.
Quelle: aws.amazon.com

AWS kündigt Amazon Verified Permissions (Vorversion) an

Heute kündigt AWS die Vorversion von Amazon Verified Permissions an – einem skalierbaren, präzisen Service für die Berechtigungsverwaltung und Autorisierungen für benutzerdefinierte Anwendungen. Mit Amazon Verified Permissions können Anwendungsentwickler ihren Endbenutzern die Verwaltung von Berechtigungen und den gemeinsamen Zugriff auf Daten ermöglichen. Anwendungsentwickler können mit Amazon Verified Permissions beispielsweise präzise Berechtigungen definieren und verwalten, um festzulegen, welche Amazon-Cognito-Benutzer Zugriff auf welche Anwendungsressourcen haben. 
Quelle: aws.amazon.com

Einführung von Amazon-EC2-R7iz-Instances

Ab heute sind speicheroptimierte, hochleistungsfähige Amazon-EC2-R7iz-Instances als Vorversion verfügbar. R7iz-Instances sind die ersten EC2-Instances mit Intel-Xeon-Scalable-Prozessoren der 4. Generation (Codename Sapphire Rapids) mit einer Turbofrequenz von bis zu 3,9 GHz für alle Kerne. Diese Instances haben die höchste Leistung pro vCPU unter den x86-basierten EC2-Instances und liefern bis zu 20 % mehr Leistung als z1d-Instances. Die Instances basieren auf dem AWS Nitro System, einer Kombination aus dedizierter Hardware und schlankem Hypervisor, der praktisch alle Rechen- und Speicherressourcen der Host-Hardware für die Instances bereitstellt und so für eine bessere Gesamtleistung und Sicherheit sorgt. R7iz-Instances sind ideal für Front-End Electronic Design Automation (EDA), relationale Datenbanken mit hohen Lizenzgebühren pro Kern, Simulationen in den Bereichen Finanzen, Versicherungsmathematik und Datenanalyse sowie für andere Workloads, die eine Kombination aus hoher Rechenleistung und großem Speicherbedarf erfordern.
Quelle: aws.amazon.com

Ankündigung der Vorversion von Amazon CloudWatch Internet Monitor

Amazon CloudWatch Internet Monitor ist eine neue Vorschaufunktion von Amazon CloudWatch, die Sie als Anwendungsentwickler und Netzwerktechniker bei der kontinuierlichen Überwachung der Internetverfügbarkeit und der Leistungskennzahlen Ihrer bei AWS gehosteten Anwendungen und Anwendungsendbenutzer unterstützt. InternetMonitor überwacht Ihre Anwendung über Amazon Virtual Private Clouds (VPCs), Verteilungen in Amazon CloudFront und Verzeichnisse in Amazon WorkSpaces.
Quelle: aws.amazon.com

Amazon RDS Optimized Writes ermöglicht einen bis zu zweifach höheren Schreibdurchsatz ohne zusätzliche Kosten

Amazon Relational Database Service (Amazon RDS) für MySQL unterstützt jetzt Amazon RDS Optimized Writes. Mit Optimized Writes können Sie den Schreibdurchsatz ohne zusätzliche Kosten bis zu 2x verbessern. Dies ist besonders nützlich für Kunden von RDS für MySQL mit schreibintensiven Datenbank-Workloads, die häufig bei Anwendungen für digitale Zahlungen, Finanzhandel und Online-Spiele auftreten.
Quelle: aws.amazon.com

Data: the Rx to faster, patient-centric clinical trials

Out of necessity, the life sciences industry has accelerated the innovation and experimentation of drug and device development. The sector, which has traditionally been slow-moving when it comes to clinical trials—for reasons ranging from regulatory, to trial recruitment, to quality control—is now looking towards cloud technology to speed up the process and find new innovative ways to support R&D. With the shift towards patient-centric care delivery and the rapid growth of health data, the case for faster digitization in life sciences has never been stronger. However, there are still a few obstacles to overcome.Innovation roadblocksThe time and costs involved in clinical trials are enormous. The average clinical trial across therapeutic areas come out to:1With these barriers to having a new drug or device approved, it’s no surprise that more than 1 in 5 clinical trials fail due to a lack of funding.2These clinical trials are also subject to stringent regulatory requirements, and the organizations conducting the study often lack efficient and secure ways to collect, store, and analyze data across trial sites. At the same time, siloed data and poor collaboration across sites make it harder to find valuable insights that could influence and accelerate outcomes.The public will likely now greet life sciences companies with less patience for decade-long drug development cycles and more demands for retail-like transparency. Because the pharmaceutical industry needs to update its processes, meeting these expectations won’t be as simple as replicating the COVID-19 vaccine model. How, then, might the industry accelerate bringing new life saving treatments to the market safely and more quickly without a public emergency? Pharma companies now have to find new and innovative ways to conduct R&D more efficiently and drive products to market faster and more efficiently.Google Cloud is empowering scientists throughout the drug discovery pipeline from target identification, to target validation, to lead identification. By combining the power of AlphaFold and Vertex AI, we are able to significantly decrease the time to process protein engineering and de-novo protein design. The value for researchers is immense and leads to: optimized compute resource time, maximized throughput, enhanced and comprehensive trackability and reproducibility. In short, we are enabling life sciences organizations to increase the velocity of protein design and engineering to revolutionize biochemical research and drug discovery.Accelerate your clinical trials in the cloudGoogle Cloud accelerates drug and device development by revolutionizing data collection, storage, and analysis to deliver life-saving treatments faster. It reduces enrollment cycle times through the expansion of clinical trial sites, research data management solutions, and Google’s cross-site collaboration solutions, includingLowering the time and cost of clinical trials.Complying with changing global regulations. Delivering seamless communication across trial sites.Increase patient participation.How Moderna boosted discovery with dataAmerican pharmaceutical company, Moderna, needed an easier and faster way to access actionable insights. Data analysis required significant manual work and led to data silos across the organization.Moderna decided to use Google Cloud for its multi-cloud data strategy and Looker for a more holistic view of its clinical trials. By integrating internal and external data sets, the company:Gained a more complete view of clinical trials.Increased scientific efficiency and collaboration.Was able to make real-time decisions to ensure trial quality.“Looker fits well with our multi-cloud philosophy because we can choose our preferred database and leverage integrations to make our data accessible and actionable.”—Dave Johnson, VP of Informatics, Data Science, and AI at ModernaTechnology can be the enabler the industry needs in the effort to meet expectations for faster and better therapies for patients while keeping the process cost-effective for drug and device makers.1. How much does a clinical trial cost?2. National Library of Medicineaside_block[StructValue([(u’title’, u’How real-world data and analytics help accelerate clinical trials’), (u’body’, <wagtail.wagtailcore.rich_text.RichText object at 0x3ed9fe0c6590>), (u’btn_text’, u’Read more on Transform with Google Cloud’), (u’href’, u’https://cloud.google.com/blog/transform/real-world-data-platform-accelerates-clinical-trails-life-sciences-healthcare-clinica’), (u’image’, None)])]
Quelle: Google Cloud Platform

Performance considerations for loading data into BigQuery

Customers have been using BigQuery for their data warehousing needs since it was introduced. Many of these  customers routinely load very large data sets into their Enterprise Data Warehouse. Whether one is doing an initial data ingestion with hundreds of TB of data or incrementally loading from systems of record, performance of bulk inserts is key to quicker insights from the data. The most common architecture for batch data loads uses Google Cloud Storage(Object storage) as the staging area for all bulk loads. All the different file formats are converted into an optimized Columnar format called ‘Capacitor’ inside BigQuery.This blog will focus on various file types for best performance. Data files that are uploaded to BigQuery, typically come in Comma Separated Values(CSV), AVRO, PARQUET, JSON, ORC formats. We are going to use two large datasets to compare and contrast each of these file formats. We will explore loading efficiencies of compressed vs. uncompressed data for each of these file formats. Data can be loaded into BigQuery using multiple tools in the GCP ecosystem. You can use the Google Cloud console, bq load command, using the BigQuery API or using the client libraries. This blog attempts to elucidate the various options for bulk data loading into BigQuery and also provides data on the performance for each file-type and loading mechanism.IntroductionThere are various factors you need to consider when loading data into BigQuery. Data file formatData compressionLevel of parallelization of data loadSchema autodetect ‘ON’ or ‘OFF’Wide tables vs narrow(fewer columns) tables.Data file formatBulk insert into BigQuery is the fastest way to insert data for speed and cost efficiency. Streaming inserts are however more efficient when you need to report on the data immediately. Today data files come in many different file types including Comma Separated(CSV), JSON, PARQUET, AVRO  to name a few. We are often asked how the file format matters and whether there are any advantages in choosing one file format over the other. CSV files (comma-separated values) contain tabular data with a header row naming the columns. When loading data one can parse the header for column names. When loading from csv files one can use the header row for schema autodetect to pick up the columns. With schema autodetect set to off, one can skip the header row and create a schema manually, using the column names in the header. CSV files can use other field separators(like ; or |) too as a separator, since many data outputs already have a comma in the data. You cannot store nested or repeated data in CSV file format.JSON (JavaScript object notation) data is stored as a key-value pair in a semi structured format. JSON is preferred as a file type because it can store data in a hierarchical format. The schemaless nature of JSON data rows gives the flexibility to evolve the schema and thus change the payload. JSON formats are user-readable. REST-based web services use json over other file types.PARQUET is a column-oriented data file format designed for efficient storage and retrieval of data.  PARQUET compression and encoding is very efficient and provides improved performance to handle complex data in bulk.AVRO: The data is stored in a binary format and the schema is stored in JSON format. This helps in minimizing the file size and maximizes efficiency. From a data loading perspective we did various tests with millions to hundreds of billions of rows with narrow to wide column data .We have done this test with a public dataset named `bigquery-public-data.samples.github_timeline` and `bigquery-public-data.wikipedia.pageviews_2022`. We used 1000 flex slots for the test and the number of loading(called PIPELINE slots) slots is limited to the number of slots you have allocated for your environment. Schema Autodetection was set to ‘NO’. For the parallelization of the data files, each file should typically be less than 256MB uncompressed for faster throughput and here is a summary of our findings:Do I compress the data? Sometimes batch files are compressed for faster network transfers to the cloud. Especially for large data files that are being transferred, it is faster to compress the data before sending over the cloud Interconnect or VPN connection. In such cases is it better to uncompress the data before loading into BigQuery? Here are the tests we did for various file types with different file sizes with different compression algorithms. Shown results are the average of five runs:How do I load the data?There are various ways to load the data into BigQuery. You can use the Google Cloud Console, command line, Client Library or use the REST API. As all these load types invoke the same API under the hood so there is no advantage of picking one way over the other. We used 1000 PIPELINE slots reservations, for doing the data loads shown above. For workloads that require predictable load times, it is imperative that one uses PIPELINE slot reservations, so that load jobs are not dependent on the vagaries of available slots in the default pool. In the real world many of our customers have multiple load jobs happening concurrently. In those cases, assigning PIPELINE slots to individual jobs has to be done carefully keeping a balance between load times and slot efficiency.Conclusion: There is no distinct advantage in loading time when the source file is in compressed format for the tests that we did. In fact for the most part uncompressed data loads in the same or faster time than compressed data. For all file types including AVRO, PARQUET and JSON it takes longer to load the data when the file is compressed. Decompression is a CPU bound activity and your mileage varies based on the amount of PIPELINE slots assigned to your load job. Data loading slots(PIPELINE slots) are different from the data querying slots. For compressed files, you should parallelize the load operation, so as to make sure that data loads are efficient. Split the data files to 256MB or less to speed up the parallelization of the data load.From a performance perspective AVRO and PARQUET files have similar load times. Fixing your schema does load the data faster than schema autodetect set to ‘ON’. Regarding ETL jobs, it is faster and simpler to do your transformation inside BigQuery using SQL, but if you have complex transformation needs that cannot be done with SQL, use Dataflow for unified batch and streaming, Dataproc for streaming based pipelines, or Cloud Data Fusion for no-code / low-code transformation needs. Wherever possible, avoid implicit/explicit data types conversions for faster load times. Please also refer to Bigquery documentation for details on data loading to BigQuery.To learn more about how Google BigQuery can help your enterprise, try out Quickstarts page hereDisclaimer: These tests were done with limited resources for BigQuery in a test environment during different times of the day with noisy neighbors, so the actual timings and the number of rows might not be reflective of your test results. The numbers provided here are for comparison sake only, so that you can choose the right file types, compression for your workload.  This testing was done with two tables, one with 199 columns (wide table) and another with 4 columns (narrow table). Your results will vary based on the datatypes, number of columns, amount of data, assignment of PIPELINE slots and various file types. We recommend that you test with your own data before coming to any conclusion.Related ArticleLearn how BI Engine enhances BigQuery query performanceThis blog explains how BI Engine enhances BigQuery query performance, different modes in BI engine and its monitoring.Read Article
Quelle: Google Cloud Platform