Shrihari Mundada 18 min

Is your Installed Base data a Mess?


Clean and unify Installed Base data with Entytle



0:00

Let me go ahead and introduce our speaker of the day.

0:04

We have Shri Moondata, who has worked with in titles

0:07

for the past eight years.

0:10

He takes care of the entire data and DevOps infrastructure

0:13

for in title.

0:14

He has been instrumental in building

0:16

an in-salt-based data studio focused on industrial data

0:20

workflows, which he's going to go ahead and jump into

0:22

with that being said, Shri, all.

0:24

But you take it away.

0:28

Thanks a lot for the introduction, Colin.

0:30

Hello, everyone.

0:31

And thank you for joining our webinar today.

0:33

I'm excited to talk to you about a common challenge

0:36

based by industrial OEMs, which is managing the install

0:40

based data effectively.

0:43

Most of you would agree to this that this is a comprehensive

0:47

process.

0:48

And with the next few minutes, I'd like to shed light on,

0:51

how do we do it differently and repeatedly, and at a larger

0:54

scale at the end of the title?

0:56

So let's start with the scenario, which many of you

1:00

might be familiar with.

1:01

Your install based data, as you see in this slide,

1:04

would be scattered across various functional systems,

1:06

like CRM systems, ERP, FSM, different spreadsheets,

1:11

or e-commerce tools, legacy systems, and so on.

1:14

And each of the systems may hold valuable information,

1:17

but accessing and unifying this data can be a daunting task.

1:21

So it's like having all the pieces of the puzzle,

1:24

but not being able to see the complete picture.

1:26

And this is what we see across every single prospect

1:30

or customer that we talk to.

1:32

On top of that, the data here would be often messy,

1:35

inconsistent formats, duplicate entries, incomplete records.

1:39

It really makes it difficult to gain accurate insights

1:42

on to the data.

1:43

It becomes a big hindrance and leads to missed revenue

1:47

opportunities.

1:49

And this problem gets further compounded by the data silos.

1:52

And it's not that companies do not

1:55

make investments in this area, but thing

1:57

is the lack of a comprehensive process

2:01

around this data-fitting cleaning notification

2:05

leads to eventually get another silo.

2:09

So what's the solution?

2:11

And where does the entire comment to picture?

2:13

So what the entire list has done is

2:15

we have created a clean and unified process

2:19

that does the data cleaning enrichment analysis

2:24

on your IB data, which keeps your IB objects at the center.

2:30

So if we look at this data journey,

2:32

so here what we are striving is to transform your raw data

2:37

into actionable insights through the entire installation

2:40

platform.

2:41

So we begin by gathering the relevant data

2:43

from different systems.

2:45

Then we move to data exploration, where

2:47

we do a lot of everything on the data quality

2:50

and the data cleaning part of it, before feeding it

2:53

to analytics pipeline, where we have built certain custom

2:56

purpose-built models centered around the IB data.

3:01

And only when we have validated those results,

3:06

practice, set them, we push it to the final destination,

3:11

which are basically actionable intelligence, which

3:15

are accessed by our customers leading

3:18

into more new opportunities for them.

3:21

Having said that, the most crucial aspect in this journey

3:24

is the data quality, as it lays the foundation

3:29

for accurate analysis and generate

3:32

reliable insights in your IB data journey.

3:35

So we'll just deep dive into the data quality piece

3:39

because it has to be really comprehensive.

3:43

So if you take a step back and understand

3:45

what are the primary objects around the IB data?

3:48

So they could be transactions, equipments,

3:52

item master, bill of materials, service data, service

3:56

contact data, warrant these location masters,

3:58

and all of these data in real life

4:00

is scattered across different systems.

4:02

So the first thing is it should be cleaned and unified

4:05

and also enriched further so that you can

4:09

trick this into a high quality install base.

4:12

And during this process, it's also

4:14

important to have proper data governance and security

4:16

practices because it's important to ensure

4:19

integrity and protection of data.

4:22

But having said that, it's a very recent and done,

4:25

basically.

4:26

As you can see in this slide, cleaning and stitching

4:29

this install base data is a comprehensive process,

4:32

but it requires a lot of investment,

4:34

diverse skillsets, multiple tools, which

4:37

is really a time consuming and expensive thing to do.

4:40

So at the entitled, we have built something called

4:43

as a part of the installation platform.

4:48

We have built something called as the entire install base

4:50

data studio, which unifies all the data from all of your

4:54

different systems into a single cohesive system.

4:58

So this ensures that we achieve high quality data,

5:03

focused around your IB use cases, and on that high quality

5:08

data, we generate C60 degree analysis

5:13

like the parts of your training, equipment

5:15

training services, revenue training, identify the health

5:17

of your customers, both at a macro and micro level,

5:20

and make specific predictions and prescriptions

5:24

around equipments, opportunities, parts, opportunities,

5:28

and so on.

5:29

So at a high level, this is what our approach is.

5:36

What I do is I directly jump into our product demo

5:39

because the proof of the pudding is in the eating,

5:42

where I would actually show, what is it that entire install

5:45

base data studio has to offer and how do we do it?

5:47

As you can see here, into the IB data studio,

5:54

we have different workspaces for different personnel.

5:56

One is data prep workspace, where we do all the data

6:00

profiling, stitching, and cleaning, and stuff.

6:03

There is the analytic and the setup, basically,

6:06

the meta data management.

6:08

In the analytics, we have descriptive, predictive,

6:10

and predictive opportunities on the clean and unified IB data

6:14

that would have done into the data prep workspace.

6:16

Once we jump into the data prep workspace,

6:18

you would see a lot of services.

6:20

All of them are like self explanatory.

6:21

So data prep is where you prep all the data,

6:24

data loader, and data load is where we ingest this data

6:28

into our target systems.

6:30

And deduplication is where we deduplicate locations,

6:34

which is a common identifier across all different systems

6:38

that we have spoken about earlier.

6:41

So on the data acquisition side, you can upload file

6:45

from your local machine, or you can import directly

6:47

from your databases, import directly from snowflake.

6:50

And once you have imported data, we run several jobs

6:54

to profile that data.

6:56

So if I jump into one of the jobs, what you can see here

6:59

is a data distribution, which helps in identifying outliers,

7:04

missing values, the patterns in your data.

7:07

Or example, if I look at the report of this particular data

7:10

center that I have, it has roughly 150,000 records.

7:14

And 30% of the data is missing.

7:18

Now, if you look at this, you have different attributes,

7:21

cardinality, or different columns, which denotes

7:24

what are the, what is the, what how many number of records

7:28

in your data are distinct?

7:29

It will buy the book, why was it your book,

7:32

so you would identify any outliers in your data.

7:35

So doing this in the initial phase helps you really identify,

7:39

helps you identify how does my data quality look like?

7:42

Is this because if take a classic case where your data

7:48

may be lying into two to three different systems,

7:49

we could unify that and bring all of that data together.

7:52

And this gives you a unified picture of your data

7:55

distribution.

7:55

And then you can take a call on how do you improve the quality

7:58

of this data?

7:59

And if it is good, you move ahead, or if not,

8:04

you can perform several elementary and complex data

8:08

operations.

8:09

Some of them are merging a column,

8:12

was splitting a column, deleting it, renaming it

8:14

for better readability.

8:16

There are no code-local operations like find and replace,

8:19

find and replace by regular expressions.

8:21

You can just find by certain pattern and then replace it

8:24

by a specific value.

8:27

There are SQL type joins, where you can do left, right,

8:31

inner join on your data.

8:34

Then so I can choose a data set and do a left, right, inner join

8:39

also out a join with any other data set.

8:41

There are no code operations like create new column of type

8:44

string number, date, Boolean.

8:46

And you can perform various operations.

8:48

So if column A is empty and column B is not empty.

8:51

So different kind of no code things.

8:54

You can define a uniqueness criteria on your data.

8:58

You can say that a combination of n number of columns defines

9:02

my record as unique.

9:04

And what it is, it would drop duplicates.

9:07

That's how you keep unique records.

9:10

Then there are other operations like SQL type group

9:13

by or SQL type ranking functions, where you can group

9:15

by certain things and then you can use certain functions.

9:19

One of the use cases is you would have

9:22

caught some initial set of data from your system.

9:24

Now you want to apply an additional data to this.

9:26

So you can union it pattern matching

9:30

barricular expressions.

9:31

So you can count and extract matches

9:33

by certain regular expressions.

9:35

And the best part is whatever you do,

9:37

you can put this as a part of the recipe so that the next time

9:42

when you get similar data, you can rerun the same recipe step

9:47

that you had written earlier.

9:50

And what you are seeing here is something

9:53

called a ribic continuum, basically, which

9:56

is a connected flow of different objects in a data set.

10:00

So I've got all of these transaction data

10:03

from different sources.

10:04

I have done a lot of cleaning that is common.

10:07

I've always objects in a single recipe

10:10

and then split into its respective destinations.

10:14

And these are the standard objects, like orders, assets,

10:17

services, service contracts, warranties, which primarily

10:20

deal with your IB use cases.

10:23

Now that this data would have been clean and enriched,

10:27

we also have a de-duplication service

10:30

where we can de-duplicate certain those locations

10:35

by whatever definition.

10:37

So you may say a combination of customer names,

10:39

the exact combination of location unique,

10:40

or just a combination of customer name

10:42

makes my accounts unique.

10:44

So as you can see here, this connects communication,

10:47

it is present across different systems.

10:49

And then we have brought all of them

10:51

as a part of the single cluster.

10:53

So this is how you de-duplicate your purchasing entities

10:58

or locations which help you unify your data

11:02

in a much better fashion.

11:05

Now that this cleaning and unification

11:08

will be done, we feed it to our analytics pipeline.

11:11

So we have a lot of services which work on this clean and unify

11:16

data and for better readability.

11:19

So I'll just show the configuration screen

11:21

where we have services both at a macro and a micro level.

11:26

At a macro level, we can generate what

11:28

is the propensity or probability that a location or transit

11:31

either part or an equipment or a service from you

11:35

in the next and number of days.

11:36

So once you train a model, you can use it for next and number

11:40

of days or you can train a new model based on data systems.

11:42

If based on the transaction history that you have got,

11:49

you can rank contacts by assigning a page

11:51

to their occurrences into different sources of the data.

11:55

And you can rank contact this, it typically

11:57

solves your whom to call problem.

11:59

Once you have opportunities created on the IV data.

12:03

We also have predictive and prescriptive opportunities

12:06

on the transaction history.

12:08

So you can generate opportunities at a micro level.

12:14

We would say that location X would transact item Y

12:18

in the next 90 days and this is the quantity and this is the

12:20

probability.

12:21

So this is and we call it as a purchase rate service.

12:25

At a micro level, we can segment your customers by usage

12:29

and the silver tellumer of parts and total number of

12:32

revenue that they would have generated.

12:34

We can also calculate the frequency, consistency and

12:37

recency of every purchasing entity or location.

12:40

What is the frequency between which they are buying?

12:41

What is the consistency?

12:43

And based on these metrics, we can also calculate

12:46

are your customers drifting away from you?

12:48

While a customer that was consistently buying from you has

12:52

that customer drifted away from you recently.

12:54

So these are the kind of information analytics that we

12:57

can generate at a macro and a micro level on that clean and

13:03

unified data.

13:04

So this is the USP of entire digital basically the studio

13:07

where we do all of this heavy lifting.

13:10

All of them under a single roof as you can see in this slide.

13:14

So you get data from different sources.

13:17

You perform, you do all the data stitching cleaning

13:22

unification.

13:23

We perform different purpose with analytics like calculating

13:27

a wallet share, a trend, propensity, and the predictive

13:32

and prescription opportunities that I spoke about a while ago.

13:35

And all of this can be viewed into a flagship product

13:40

which is the entire inside which gives you a 360 degree

13:42

visibility on your installments data where you can do

13:46

search or filtering or prioritize your hunting list.

13:50

You can also push this data directly into your CRM or into

13:54

data lakes.

13:54

We have customers who have given us access to their CRM

13:59

systems where we directly push it to their sales force or any

14:03

other system that would have configured.

14:07

We also have a lot of open APIs wherein our customers access

14:11

this clean and unified data via APIs.

14:15

And gain the value or basically derive maximum value on the

14:19

skin and unified data.

14:22

And one such example is how effortlessly we can integrate

14:26

this data into a sales force CRM.

14:29

So what you see, what you heard from me in the previous slides

14:33

before the past few minutes is all your major IB objects,

14:39

like equipments, parts, services, service contacts.

14:42

Once you have clean, enriched and unified that into the

14:45

entire install based intelligence platform, all of them can

14:48

be pulled seamlessly into your sales force account.

14:53

Thereby giving you a 360 degree view under a single roof

14:57

and giving you more than any opportunities and you can do

15:01

and there are endless possibilities to do things around

15:04

your install based data.

15:07

So this is what I had.

15:09

But today in the next webinar, we would be talking about

15:13

integration with sales force and also how

15:16

entitled analytics help our customers generate more revenue

15:21

by means of more predictive and prescriptive opportunities.

15:26

So with this, I would like to leave it to the audience for

15:30

any questions.

15:31

Yeah, we had a couple of questions come in.

15:37

So I'm going to ask you a question.

15:39

So I'm going to ask you a question.

15:41

So I'm going to ask you a question.

15:43

So I'm going to ask you a question.

15:45

So I'm going to ask you a question.

15:47

So I'm going to ask you a question.

15:49

So I'm going to ask you a question.

15:51

So I'm going to ask you a question.

15:53

So I'm going to ask you a question.

15:55

So I'm going to ask you a question.

15:57

So I'm going to ask you a question.

15:59

So I'm going to ask you a question.

16:01

So I'm going to ask you a question.

16:03

So I'm going to ask you a question.

16:05

So I'm going to ask you a question.

16:07

So I'm going to ask you a question.

16:09

So I'm going to ask you a question.

16:11

So I'm going to ask you a question.

16:13

So I'm going to ask you a question.

16:15

So I'm going to ask you a question.

16:17

So I'm going to ask you a question.

16:19

So I'm going to ask you a question.

16:21

So I'm going to ask you a question.

16:23

So I'm going to ask you a question.

16:25

So I'm going to ask you a question.

16:27

So I'm going to ask you a question.

16:29

So I'm going to ask you a question.

16:31

So I'm going to ask you a question.

16:33

And then we have another question here.

16:35

How easy is it to integrate and title with sales force?

16:39

And what does the add value it can give?

16:43

Yeah, so.

16:47

So it's a again a very good question.

16:49

And that's as you can see on my slide.

16:51

This is what you see once you have clean and identify data into your sales for

16:56

the account.

16:57

Where under that.

16:58

So under this I be intelligent staff just created by entitled.

17:01

You would see a 360 degree view of all your IB objects under a single tap.

17:09

So it's once the data is cleaned and unified.

17:13

The next step is pretty easy.

17:17

And we are basically we are very easy for our customers to have all of the data

17:23

under a single.

17:25

Right, they're keeping up with their current work flows that they have or the

17:28

current text set that they use.

17:29

But have the capability of expedited account research done all in front of them

17:37

Okay, that's all the questions that we had to come through today.

17:43

If anybody has any sort of follow up questions, please feel free to reach out

17:47

to us.

17:49

And we can get you connected to the right party here at entitled.

17:56

Right, with that being said, thanks everybody.

17:59

We look forward to seeing you in the next event.

18:03

Thank you.