Using ML and Azure to improve Customer Lifetime Value: On-Demand Webinar and FAQ Now Available!

Companies are generating value from Customer Lifetime Value in many ways, but the two most common are: Improved revenue.

80% of your company’s future revenue will come from just 20% of your existing customers.

Establishing a CLV metric enables you to focus your promotions on this high-revenue pool, and attract more customers like your high-value customers.

The probability of selling to an existing customer is 60-70%.

The probability of selling to a new prospect is 5-20%.

Improved profitability.

Reducing churn by 5% can increase profits by 25-125%.

Reducing retention costs.

Your Marketing team should be able to provide a metric on the average cost to acquire a new customer.

The retention benefit is the difference between the acquisition cost and retention cost, multiplied by the number of customers being retained.

It costs 5 times more to acquire new customers than it does to keep current ones.

You should expect to see an improvement in your Cost to Serve as a result of CLV.

Focusing your conversion and retention programs on the most profitable customers will have a direct impact on Cost to Serve, but this number is difficult to estimate without the CLV first being in place.

These numbers scale based on the number of customers you have.

BlueGranite can work with you to build your business case.

Q: In the demo, what’s the relationship between the total sales amount and the probability of churn?.Is one a function of the other?.The relationship is not used in my definition of churn.

But there are functions in Spark, where you could establish the relationship.

You can do more model explainability.

You can define churn however you want.

We have defined churns as  “has the customer purchased anything in 10 months are not?” Q: How does a developer determine what type of worker is needed for their requirement?.RAM/CPU needs etc.

?.The way Apache SparkTM hands out things to do for the cluster is based on tasks.

If a node in the cluster has 4 cores, then it can handle 4 tasks simultaneously.

Generally, the more cores you have the more tasks you can run simultaneously.

Memory is important when you are dealing with a large dataset.

Spark runs in memory.

The larger the data set, the more memory you need.

I work with genomics data sets that are huge, I like to work with memory-optimized machines that have a lot more memory per core.

However, the default machine types are good enough for more use cases.

  Also, if you chose to work on use cases that involve software such as TensorFlow, Keras and Horovod that require GPUs, you would have to use specific machines that have GPUs.

Q: Is there any function for getting names for the feature importance vectors for categorical columns?.I don’t have the code ready here, but I have had the opportunity to do this before.

What happens with the categorical columns is it dummy codes it using onehotencoding.

It turns one categorical column into multiple columns that are numerical.

There is a way to then determine 1 means red and 2 means blue.

There is a way to join them back together.

Q: How can I contact BlueGranite?.The easiest way to reach us is through our website’s contact-us form: https://www.

blue-granite.

com/contact-us Q: How can BlueGranite help get us started on Databricks?.We take a ‘start small, grow big’ approach.

Initially, we’ll want to understand your use-cases and the success factors you’re trying to achieve.

Then we can help get you up and running fast through a combination of education and collaborative engagements with your team.

We can help educate you or assist in your platform evaluation through workshops, hands-on training, envisioning sessions, or a quick-start engagement where we’ll implement one of your use-cases from end-to-end.

  Try Databricks for free.

Get started today.. More details

Leave a Reply