Wednesday, October 11, 2023
HomeBig DataBe a part of AWS Databricks clients at Knowledge + AI Summit...

Be a part of AWS Databricks clients at Knowledge + AI Summit 2023


It is a collaborative publish from Databricks and Amazon Net Companies (AWS). We thank Venkat Viswanathan, Knowledge and Analytics Technique Chief, Accomplice Options at AWS, for his contributions.

 

Knowledge + AI Summit 2023: Register now to affix this in-person and digital occasion June 26-29 and study from the worldwide knowledge neighborhood.

Amazon Net Companies (AWS) is a Platinum Sponsor of Knowledge + AI Summit 2023, the premier occasion for the worldwide knowledge neighborhood. Be a part of this occasion and study from joint Databricks and AWS clients like Labcorp, Conde Nast, Grammarly, Vizio, NTT Knowledge, Impetus, Amgen, and YipitData who’ve efficiently leveraged the Databricks Lakehouse Platform for his or her enterprise, bringing collectively knowledge, AI and analytics on one frequent platform.

At Knowledge + AI Summit, Databricks and AWS clients will take the stage for periods that will help you see how they achieved enterprise outcomes utilizing the Databricks on AWS Lakehouse. Attendees can have the chance to listen to knowledge leaders from Labcorp on Tuesday, June twenty seventh, then be a part of Grammarly, Vizio, NTT Knowledge, Impetus, and Amgen on Wednesday, June twenty eighth and Conde Nast and YipitData on Thursday, June twenty ninth. At Knowledge + AI Summit, study concerning the newest improvements and applied sciences and listen to thought-provoking panel discussions together with the power for networking alternatives the place you’ll be able to join with different knowledge professionals in your business.

AWS might be showcasing how you can make the most of AWS native companies with Databricks at each their AWS sales space and Demo Stations:

In Demo Station 1 – AWS might be showcasing how clients can leverage AWS native companies together with AWS Glue, Amazon Athena, Amazon Kinesis, Amazon S3, to investigate Delta Lake.

  • Databricks Lakehouse platform with AWS Glue, Amazon Athena, and Amazon S3
  • AWS IoT Hub, Amazon Kinesis Knowledge Streams, Databricks Lakehouse platform, Amazon S3 (probably extending to Quicksight)
  • SageMaker JumpStart, Databricks created Dolly 2.0 and different open supply LLMs, Amazon OpenSearch
  • SageMaker Knowledge Wrangler and Databricks Lakehouse platform

In Demo Station 2 – AWS will completely show Amazon Quicksight integration with Databricks Lakehouse platform

  • Databricks Lakehouse platform, Amazon QuickSight, Amazon QuickSight Q

Please cease by the Demo Stations and the AWS sales space to study extra about Databricks on AWS, meet the AWS crew, and ask questions.

The periods beneath are a information for everybody taken with Databricks on AWS and span a variety of subjects — from knowledge observability, to reducing complete value of possession, to demand forecasting and safe knowledge sharing. When you’ve got questions on Databricks on AWS or service integrations, join with Databricks on AWS Options Architects at Knowledge + AI Summit.

Databricks on AWS buyer breakout periods

Labcorp Knowledge Platform Journey: From Choice to Go-Dwell in Six Months

Tuesday, June 27 @3:00 PM

Be a part of this session to study concerning the Labcorp knowledge platform transformation from on-premises Hadoop to AWS Databricks Lakehouse. We’ll share greatest practices and classes discovered from cloud-native knowledge platform choice, implementation, and migration from Hadoop (inside six months) with Unity Catalog.

We’ll share steps taken to retire a number of legacy on-premises applied sciences and leverage Databricks native options like Spark streaming, workflows, job swimming pools, cluster insurance policies and Spark JDBC inside Databricks platform. Classes discovered in Implementing Unity Catalog and constructing a safety and governance mannequin that scales throughout purposes. We’ll present demos that stroll you thru batch frameworks, streaming frameworks, knowledge examine instruments used throughout a number of purposes to enhance knowledge high quality and pace of supply.

Uncover how now we have improved operational effectivity, resiliency and lowered TCO, and the way we scaled constructing workspaces and related cloud infrastructure utilizing Terraform supplier.

Be taught extra

How Comcast Effectv Drives Knowledge Observability with Databricks and Monte Carlo

Tuesday, June 27 @4:00 PM

Comcast Effectv, the two,000-employee promoting wing of Comcast, America’s largest telecommunications firm, gives customized video advert options powered by aggregated viewership knowledge. As a world expertise and media firm connecting tens of millions of consumers to customized experiences and processing billions of transactions, Comcast Effectv was challenged with dealing with large a great deal of knowledge, monitoring a whole lot of knowledge pipelines, and managing well timed coordination throughout knowledge groups.

On this session, we’ll focus on Comcast Effectv’s journey to constructing a extra scalable, dependable lakehouse and driving knowledge observability at scale with Monte Carlo. This has enabled Effectv to have a single pane of glass view of their whole knowledge surroundings to make sure shopper knowledge belief throughout their whole AWS, Databricks, and Looker surroundings.

Be taught extra

Deep Dive Into Grammarly’s Knowledge Platform

Wednesday, June 28 @11:30 AM

Grammarly helps 30 million individuals and 50,000 groups to speak extra successfully. Utilizing the Databricks Lakehouse Platform, we will quickly ingest, remodel, mixture, and question advanced knowledge units from an ecosystem of sources, all ruled by Unity Catalog. This session will overview Grammarly’s knowledge platform and the selections that formed the implementation. We’ll dive deep into some architectural challenges the Grammarly Knowledge Platform crew overcame as we developed a self-service framework for incremental occasion processing.

Our funding within the lakehouse and Unity Catalog has dramatically improved the pace of our knowledge worth chain: making 5 billion occasions (ingested, aggregated, de-identified, and ruled) obtainable to stakeholders (knowledge scientists, enterprise analysts, gross sales, advertising and marketing) and downstream companies (function retailer, reporting/dashboards, buyer assist, operations) obtainable inside 15. Because of this, now we have improved our question value efficiency (110% quicker at 10% the price) in comparison with our legacy system on AWS EMR.

I’ll share structure diagrams, their implications at scale, code samples, and issues solved and to be solved in a technology-focused dialogue about Grammarly’s iterative lakehouse knowledge platform.

Be taught extra

Having Your Cake and Consuming it Too: How Vizio Constructed a Subsequent-Era ACR Knowledge Platform Whereas Decreasing TCO

Wednesday, June 28 @1:30 PM

As the highest producer of good TVs, Vizio makes use of TV knowledge to drive its enterprise and supply clients with greatest digital experiences. Our firm’s mission is to repeatedly enhance the viewing expertise for our clients, which is why we developed our award-winning automated content material recognition (ACR) platform. Once we first constructed our knowledge platform nearly ten years in the past, there was no single platform to run a knowledge as a service enterprise, so we obtained inventive and constructed our personal by stitching collectively completely different AWS companies and a knowledge warehouse. As our enterprise wants and knowledge volumes have grown exponentially over time, we made the strategic resolution to replatform on Databricks Lakehouse, because it was the one platform that might fulfill all our wants out-of-the-box akin to BI analytics, real-time streaming, and AI/ML. Now the Lakehouse is our sole supply of reality for all analytics and machine studying initiatives. The technical worth of the Databricks Lakehouse platform, akin to conventional knowledge warehousing low-latency question processing with advanced joins due to Photon to utilizing Apache Spark™ structured streaming; analytics and mannequin serving, might be lined on this session as we speak about our path to the Lakehouse.

Be taught extra

Why a Main Japanese Monetary Establishment Selected Databricks to Speed up its Knowledge and AI-Pushed Journey

Wednesday, June 28 @2:30 PM

On this session, we’ll introduce a case examine of migrating the Japanese largest knowledge evaluation platform to Databricks.

NTT DATA is without doubt one of the largest system integrators in Japan. Within the Japanese market, many firms are engaged on BI, and we are actually within the section of utilizing AI. Our crew gives options that present knowledge analytics infrastructure to drive the democratization of knowledge and AI for main Japanese firms.

The shopper on this case examine is without doubt one of the largest monetary establishments in Japan. This mission has the next traits:

As a monetary establishment, safety necessities are very strict.

Since it’s used company-wide, together with group firms, it’s essential to assist varied use circumstances.

We began working a knowledge evaluation platform on AWS in 2017. Over the following 5 years, we leveraged AWS-managed companies akin to Amazon EMR, Amazon Athena, and Amazon SageMaker to modernize our structure. Within the close to future, in an effort to promote the use circumstances of AI in addition to BI extra effectively, now we have begun to contemplate upgrading to a platform that realizes each BI and AI. This session will cowl:

Challenges in creating AI on a DWH-based knowledge evaluation platform and why a knowledge lakehouse is the only option.

Analyzing the structure of a platform that helps each AI and BI use circumstances.

On this case examine, we’ll introduce the outcomes of a comparative examine of a proposal primarily based on Databricks, a proposal primarily based on Snowflake, and a proposal combining Snowflake and Databricks. This session is really helpful for many who wish to speed up their enterprise by using AI in addition to BI.

Be taught extra

Impetus | Accelerating ADP’s Enterprise Transformation with a Fashionable Enterprise Knowledge Platform

Wednesday, June 28 @2:30 PM

Be taught How ADP’s Enterprise Knowledge Platform Is used to drive direct monetization alternatives, differentiate its options, and enhance operations. ADP is constantly trying to find methods to extend innovation velocity, time-to-market, and enhance the general enterprise effectivity. Making knowledge and instruments obtainable to groups throughout the enterprise whereas decreasing knowledge governance threat is the important thing to creating progress on all fronts. Find out about ADP’s enterprise knowledge platform that created a single supply of reality with centralized instruments, knowledge property, and companies. It allowed groups to innovate and acquire insights by leveraging cross-enterprise knowledge and central machine studying operations.

Discover how ADP accelerated creation of the info platform on Databricks and AWS, obtain quicker enterprise outcomes, and enhance general enterprise operations. The session can even cowl how ADP considerably lowered its knowledge governance threat, elevated the model by amplifying knowledge and insights as a differentiator, elevated knowledge monetization, and leveraged knowledge to drive human capital administration differentiation.

Be taught extra

From Insights to Suggestions: How SkyWatch Predicts Demand for Satellite tv for pc Imagery Utilizing Databricks

Wednesday, June 28 @3:30 PM

SkyWatch is on a mission to democratize earth remark knowledge and make it easy for anybody to make use of.

On this session, you’ll study how SkyWatch aggregates demand indicators for the EO market and turns them into monetizable suggestions for satellite tv for pc operators. Skywatch’s Knowledge & Platform Engineer, Aayush will share how the crew constructed a serverless structure that synthesizes buyer requests for satellite tv for pc photos and identifies geographic areas with excessive demand, serving to satellite tv for pc operators maximize income and satisfying a broad vary of EO knowledge hungry customers.

This session will cowl:

  • Challenges with Achievement in Earth Statement ecosystem
  • Processing massive scale GeoSpatial Knowledge with Databricks
  • Databricks in-built H3 features
  • Delta Lake to effectively retailer knowledge leveraging optimization methods like Z-Ordering
  • Knowledge LakeHouse Structure with Serverless SQL Endpoints and AWS Step Features
  • Constructing Tasking Suggestions for Satellite tv for pc Operators

Be taught extra

Enabling Knowledge Governance at Enterprise Scale Utilizing Unity Catalog

Wednesday, June 28 @3:30 PM

Amgen has invested in constructing fashionable, cloud-native enterprise knowledge and analytics platforms over the previous few years with a concentrate on tech rationalization, knowledge democratization, general consumer expertise, improve reusability, and cost-effectiveness. Certainly one of these platforms is our Enterprise Knowledge Cloth which focuses on pulling in knowledge throughout features and offering capabilities to combine and join the info and govern entry. For some time, now we have been attempting to arrange sturdy knowledge governance capabilities that are easy, but simple to handle by Databricks. There have been a number of instruments available in the market that solved a number of rapid wants, however none solved the issue holistically. To be used circumstances like sustaining governance on extremely restricted knowledge domains like Finance and HR, a long-term answer native to Databricks and addressing the beneath limitations was deemed essential:

The best way these instruments had been arrange, allowed the overriding of some safety insurance policies

  • Instruments weren’t UpToDate with the newest DBR runtime
  • Complexity of implementing fine-grained safety
  • Coverage administration – AWS IAM + In device insurance policies

To deal with these challenges, and for large-scale enterprise adoption of our governance functionality, we began engaged on UC integration with our governance processes. With an purpose to comprehend the next tech advantages:

  • Unbiased of Databricks runtime
  • Simple fine-grained entry management
  • Eradicated administration of IAM roles
  • Dynamic entry management utilizing UC and dynamic views

Immediately, utilizing UC, now we have to implement fine-grained entry management & governance for the restricted knowledge of Amgen. We’re within the technique of devising a sensible migration & change administration technique throughout the enterprise.

Be taught extra

Activate Your Lakehouse with Unity Catalog

Thursday, June 29 @1:30 PM

Constructing a lakehouse is simple at this time due to many open supply applied sciences and Databricks. Nonetheless, it may be taxing to extract worth from lakehouses as they develop with out sturdy knowledge operations. Be a part of us to find out how YipitData makes use of the Unity Catalog to streamline knowledge operations and uncover greatest practices to scale your individual Lakehouse. At YipitData, our 15+ petabyte Lakehouse is a self-service knowledge platform constructed with Databricks and AWS, supporting analytics for a knowledge crew of over 250. We’ll share how leveraging Unity Catalog accelerates our mission to assist monetary establishments and companies leverage various knowledge by:

  • Enabling purchasers to universally entry our knowledge by a spectrum of channels, together with Sigma, Delta Sharing, and a number of clouds
  • Fostering collaboration throughout inside groups utilizing a knowledge mesh paradigm that yields wealthy insights
  • Strengthening the integrity and safety of knowledge property by ACLs, knowledge lineage, audit logs, and additional isolation of AWS sources
  • Lowering the price of massive tables with out downtime by automated knowledge expiration and ETL optimizations on managed delta tables

By way of our migration to Unity Catalog, now we have gained ways and philosophies to seamlessly circulate our knowledge property internally and externally. Knowledge platforms have to be value-generating, safe, and cost-effective in at this time’s world. We’re excited to share how Unity Catalog delivers on this and helps you get essentially the most out of your lakehouse.

Be taught extra

Knowledge Globalization at Conde Nast Utilizing Delta Sharing

Thursday, June 29 @1:30 PM

Databricks has been a necessary a part of the Conde Nast structure for the previous few years. Previous to constructing our centralized knowledge platform, “evergreen,” we had related challenges as many different organizations; siloed knowledge, duplicated efforts for engineers, and a scarcity of collaboration between knowledge groups. These issues led to distrust in knowledge units and made it troublesome to scale to fulfill the strategic globalization plan we had for Conde Nast.

Over the previous few years now we have been extraordinarily profitable in constructing a centralized knowledge platform on Databricks in AWS, totally embracing the lakehouse imaginative and prescient from end-to-end. Now, our analysts and entrepreneurs can derive the identical insights from one dataset and knowledge scientists can use the identical datasets to be used circumstances akin to personalization, subscriber propensity fashions, churn fashions and on-site suggestions for our iconic manufacturers.

On this session, we’ll focus on how we plan to include Unity Catalog and Delta Sharing as the following section of our globalization mission. The evergreen platform has grow to be the worldwide customary for knowledge processing and analytics at Conde. To be able to handle the worldwide knowledge and adjust to GDPR necessities, we’d like to ensure knowledge is processed within the acceptable area and PII knowledge is dealt with appropriately. On the similar time, we have to have a world view of the info to permit us to make enterprise selections on the international degree. We’ll speak about how delta sharing permits us a easy, safe approach to share de-identified datasets throughout areas in an effort to make these strategic enterprise selections, whereas complying with safety necessities. Moreover, we’ll focus on how Unity Catalog permits us to safe, govern and audit these datasets in a straightforward and scalable method.

Be taught extra

Databricks on AWS breakout periods

AWS | Actual Time Streaming Knowledge Processing and Visualization Utilizing Databricks DLT, Amazon Kinesis, and Amazon QuickSight

Wednesday, June 28 @11:30 AM

Amazon Kinesis Knowledge Analytics is a managed service that may seize streaming knowledge from IoT units. Databricks Lakehouse platform gives ease of processing streaming and batch knowledge utilizing Delta Dwell Tables. Amazon Quicksight with highly effective visualization capabilities can gives varied superior visualization capabilities with direct integration with Databricks. Combining these companies, clients can seize, course of, and visualize knowledge from a whole lot and 1000’s of IoT sensors with ease.

Be taught extra

AWS | Constructing Generative AI Resolution Utilizing Open Supply Databricks Dolly 2.0 on Amazon SageMaker

Wednesday, June 28 @2:30 PM

Create a customized chat-based answer to question and summarize your knowledge inside your VPC utilizing Dolly 2.0 and Amazon SageMaker. On this discuss, you’ll study Dolly 2.0, Databricks, state-of-the-art, open supply, LLM, obtainable for industrial and Amazon SageMaker, AWS’s premiere toolkit for ML builders. You’ll discover ways to deploy and customise fashions to reference your knowledge utilizing retrieval augmented technology (RAG) and extra advantageous tuning methods…all utilizing open-source elements obtainable at this time.

Be taught extra

Processing Delta Lake Tables on AWS Utilizing AWS Glue, Amazon Athena, and Amazon Redshift

Thursday, June 29 @1:30 PM

Delta Lake is an open supply mission that helps implement fashionable knowledge lake architectures generally constructed on cloud storages. With Delta Lake, you’ll be able to obtain ACID transactions, time journey queries, CDC, and different frequent use circumstances on the cloud.

There are loads of use circumstances of Delta tables on AWS. AWS has invested so much on this expertise, and now Delta Lake is accessible with a number of AWS companies, akin to AWS Glue Spark jobs, Amazon EMR, Amazon Athena, and Amazon Redshift Spectrum. AWS Glue is a serverless, scalable knowledge integration service that makes it simpler to find, put together, transfer, and combine knowledge from a number of sources. With AWS Glue, you’ll be able to simply ingest knowledge from a number of knowledge sources akin to on-prem databases, Amazon RDS, DynamoDB, MongoDB into Delta Lake on Amazon S3 even with out experience in coding.

This session will show how you can get began with processing Delta Lake tables on Amazon S3 utilizing AWS Glue, and querying from Amazon Athena, and Amazon Redshift. The session additionally covers latest AWS service updates associated to Delta Lake.

Be taught extra

Databricks-led periods

Utilizing DMS and DLT for Change Knowledge Seize

Tuesday, June 27 @2:00 PM

Bringing in Relational Knowledge Retailer (RDS) knowledge into your knowledge lake is a important and essential course of to facilitate use circumstances. By leveraging AWS Database Migration Companies (DMS) and Databricks Delta Dwell Tables (DLT) we will simplify change knowledge seize out of your RDS. On this discuss, we might be breaking down this advanced course of by discussing the basics and greatest practices. There can even be a demo the place we convey this all collectively

Be taught extra

Learnings From the Discipline: Migration From Oracle DW and IBM DataStage to Databricks on AWS

Wednesday, June 28 @2:30 PM

Legacy knowledge warehouses are pricey to take care of, unscalable and can’t ship on knowledge science, ML and real-time analytics use circumstances. Migrating out of your enterprise knowledge warehouse to Databricks permits you to scale as your small business wants develop and speed up innovation by working all of your knowledge, analytics and AI workloads on a single unified knowledge platform.

Within the first a part of this session we’ll information you thru the well-designed course of and instruments that may enable you from the evaluation section to the precise implementation of an EDW migration mission. Additionally, we’ll deal with methods to transform PL/SQL proprietary code to an open customary python code and make the most of PySpark for ETL workloads and Databricks SQL’s knowledge analytics workload energy.

The second a part of this session might be primarily based on an EDW migration mission of SNCF (French nationwide railways); one of many main enterprise clients of Databricks in France. Databricks partnered with SNCF emigrate its actual property entity from Oracle DW and IBM DataStage to Databricks on AWS. We’ll stroll you thru the client context, urgency to migration, challenges, goal structure, nitty-gritty particulars of implementation, greatest practices, suggestions, and learnings in an effort to execute a profitable migration mission in a really accelerated timeframe.

Be taught extra

Embracing the Way forward for Knowledge Engineering: The Serverless, Actual-Time Lakehouse in Motion

Wednesday, June 28 @2:30 PM

As we enterprise into the way forward for knowledge engineering, streaming and serverless applied sciences take middle stage. On this enjoyable, hands-on, in-depth and interactive session you’ll be able to study concerning the essence of future knowledge engineering at this time.

We’ll sort out the problem of processing streaming occasions constantly created by a whole lot of sensors within the convention room from a serverless net app (convey your telephone and be part of the demo). The main target is on the system structure, the concerned merchandise and the answer they supply. Which Databricks product, functionality and settings might be most helpful for our state of affairs? What does streaming actually imply and why does it make our life simpler? What are the precise advantages of serverless and the way “serverless” is a specific answer?

Leveraging the facility of the Databricks Lakehouse Platform, I’ll show how you can create a streaming knowledge pipeline with Delta Dwell Tables ingesting knowledge from AWS Kinesis. Additional, I will make the most of superior Databricks workflows triggers for environment friendly orchestration and real-time alerts feeding right into a real-time dashboard. And since I do not need you to depart with empty palms – I’ll use Delta Sharing to share the outcomes of the demo we constructed with each participant within the room. Be a part of me on this hands-on exploration of cutting-edge knowledge engineering methods and witness the long run in motion.

Be taught extra

Seven Issues You Did not Know You Can Do with Databricks Workflows

Wednesday, June 28 @3:30 PM

Databricks workflows has come a great distance because the preliminary days of orchestrating easy notebooks and jar/wheel recordsdata. Now we will orchestrate multi-task jobs and create a series of duties with lineage and DAG with both fan-in or fan-out amongst a number of different patterns and even run one other Databricks job instantly inside one other job.

Databricks workflows takes its tag: “orchestrate something anyplace” fairly severely and is a very fully-managed, cloud-native orchestrator to orchestrate numerous workloads like Delta Dwell Tables, SQL, Notebooks, Jars, Python Wheels, dbt, SQL, Apache Spark™, ML pipelines with glorious monitoring, alerting and observability capabilities as nicely. Mainly, it’s a one-stop product for all orchestration wants for an environment friendly lakehouse. And what’s even higher is, it offers full flexibility of working your jobs in a cloud-agnostic and cloud-independent approach and is accessible throughout AWS, Azure and GCP.

On this session, we’ll focus on and deep dive on a number of the very attention-grabbing options and can showcase end-to-end demos of the options which is able to can help you take full benefit of Databricks workflows for orchestrating the lakehouse.

Be taught extra

Register now to affix this free digital occasion and be a part of the info and AI neighborhood. Find out how firms are efficiently constructing their Lakehouse structure with Databricks on AWS to create a easy, open and collaborative knowledge platform. Get began utilizing Databricks with a free trial on AWS Market or swing by the AWS sales space to study extra a couple of particular promotion. Be taught extra about Databricks on AWS.



Supply hyperlink

RELATED ARTICLES

LEAVE A REPLY

Please enter your comment!
Please enter your name here

- Advertisment -
Google search engine

Most Popular

Recent Comments