- TypeConference
- Location Bangalore, Karnataka, India
- Date 25-04-2019
Computer Science
Telecommunications
Information Technology
Computer/Technology
As data continues to grow in both volume and formats across multiple deployments, performing analytics has become more complicated. By 2019, 75% of analytic solutions will incorporate 10 or more external data sources. Organizations able to glean insights from this diverse data set will have competitive advantages, from deeper understanding of customers, better responsiveness to trends, and more efficient operations, to name just a few.
This reality of data diversity has given rise to the “data lake”-a data management architecture that allows organizations store and analyze a wide variety structured and unstructured data.
A data lake is a method of data storage. What makes this approach unique is that all of the data is stored in its native format. This means that data in the lake might include everything from highly structured files to completely unstructured data such as videos, emails and images.
In addition, it is not only IT that is now integrating data. Business users are also getting involved with new self-service data preparation tools. The question is, is this the only way to manage data? Is there another level that we can get reach to allow us to more easily manage and govern data across an increasingly complex data landscape?
This seminar/Conference looks at the challenges faced by companies trying to deal with an exploding number of data sources, collecting data in multiple data stores (Cloud and on-premises), multiple analytical systems and at the requirements to be able to define, govern, manage and share trusted high quality information in a distributed and hybrid computing environment.