I recently caught up with Atif Kureishy, Global VP of Emerging Practices at Teradata, during the 2019 edition of the NVIDIA GPU Technology Conference, to get a deep dive update for how Teradata is advancing into the fields of AI and deep learning. Based in San Diego, Atif specializes in enabling clients across all major industry verticals, through strategic partnerships to deliver complex analytical solutions built on machine and deep learning. His teams are trusted advisors to the world’s most innovative companies to develop next-generation capabilities for strategic data-driven outcomes in areas of artificial intelligence, deep learning and data science.
insideBIGDATA: Can you give us a brief overview of
your role with Teradata and the company’s current direction?
Atif Kureishy: I’ve been at Teradata for two and a
half years now. In that role of promoting AI and deep learning, my teams do
three things. Number one, we work with Fortune 500 companies to advance their deep
learning capabilities. We also have a separate ML team. My team is focused on
deep learning, leveraging a lot of open source, building solutions to solve
problems and drive what we call “answers or outcomes.” The second thing we do
is we take all of that experience, lessons learned, frameworks and toolkits, approaches,
and IP, and work with our product teams to say, “Look, here’s what we did.
Here’s what was successful. Here’s what’s not. Here’s what’s hype. Here’s
what’s in the applied research community. Let’s consider how we would bring all
that into the technology portfolio of Teradata and shape the roadmaps and shape
where we’re going in the future. As part of that we’re looking to GPU
accelerate and provide deep learning capabilities into next year. And item
number three is working with media, industry analysts, with the community,
conferences to help people understand where the hype is, where the reality is,
and that takes a big portion of our time. We meet with customers to help them
understand what the state-of-the-art is, where the opportunities are, and how
to get there.
So that’s
generally me and my team. I have a global remit, so I see all geographies.
China, Japan, Europe, North America are probably the areas that we focus on
most with Singapore, Australia a little bit, but really, those are the
territories that have the most investment and spend of our customers.
The industry
that we focus on most, not surprisingly, is financial services, in particular
banking. But we also deal with telecommunications, media/entertainment, retail,
and manufacturing. Those are always the early adopters of tech, so that’s what
we’re seeing.
insideBIGDATA: How is Teradata changing their
message in light of changes in the tech industry these days?
Atif Kureishy: The reason I came to Teradata is
because the company was in the midst of a strategic transformation and I was enamored
with that transformation. We’ve done MPP really well, serving the largest
enterprises. Over time those customers get acclimated and they say, “Well,
it’s great. We keep the lights on, but I want to do more with this data. You’re
helping me with the BI reporting, but increasingly I want to apply more
advanced analytics, and I want to get into the predictions.” So Teradata
realized it needed to be more of an analytics platform, and enable our
customers to do modern data science and modern analytics.
To do this
we needed to bring in new talent, and also transform the product. It’s more
than just a SQL-based engine. It’s appreciating that you have different types
of analytics and computation that you need to apply on that data, relational
data. That’s semi-structured, unstructured, image, voice, etc. Increasingly our
customers want to apply multimodal types of analysis. Teradata is now on this pivot
to approach this transformation.
Recognizing
that if you’re moving to get out of the SQL data infrastructure game, then the
buyers change and the marketing and the go-to market changes. Who has the
dollars for AI? It’s really the line-of-business that has the dollars for AI
along with the AI agenda. In order to engage with the senior executive in the
business, you try to help that leader with any number of outcomes, like people
prediction or anomaly detection or yield optimization. You need to speak their
language, understand their business, but ultimately bring data together and
apply machine and deep learning to those problems. Narrowly, I will say I am
focused on specifically deep learning because we focus on new value creation. If
you can apply deep learning in the enterprise and solve the problem of AI explainability
then you can do things you’ve never been able to do before. For instance, in
working with a very large manufacturer, doing high pressure hose fabrication,
their default rates of scrap was off the charts. They had something like 30% of
what they manufactured end up as scrap. When we took a closer look at that, we
found they had sensors that were misaligned and as a result they had many false
positives, and they had a lot of teardown from the quality teams for
high-pressure hoses that were completely fine. So if you can take process data,
and sensor data, and apply neural networks to increase the accuracy of when you
predict a defect, the net effect is hundreds of millions of dollars.
In another
instance, you can start to understand customer behaviors when they come into a
store by using at computer vision techniques. This is what Amazon is doing with
their cashier-less stores. You have sensors in really high definition cameras,
and you can start to track how customers traverse through the store, and start
to appreciate how long they dwell and queue and different things like that. You
can optimize the layout of your store. You can put digital signage in the right
places. You can optimize your staff deployment. We’ve been working with the
largest retailers for a very long time and many are interested in this
technology. How do we change the game and allow them to do this? It’s all
predicated on data, but obviously, you need that data, and you need to analyze
that data. So that’s why Teradata recommends that they need to get into that
sort of analytic space and into the predictive space, hence the transformation.
insideBIGDATA: That’s a pretty big refocus. What was the time frame of this pivot for Teradata?
Atif Kureishy: It’s been over the last three years. We’ve
been internally focused on it, but if you’ve seen the sort of rebranding and
refresh of our go-to-market, we’re focused on pervasive data intelligence. Let
me break down those three words. “Pervasive” in the sense of you need to be
able to process all this different types of machine data, log data, structure
data, curated data, etc. and process it where it is – in the cloud, on-prem, in
object stores, in relation stores. Increasingly if you do analytics on samples
of data, you don’t really get the full view. Scale becomes a big issue and Teradata
has always been about performance at scale. The second word, “data,” is our
legacy. Finally, “intelligence” is the appreciation of artificial intelligence,
and the way of prediction and better insights and understanding is on that data,
at scale, everywhere.
So in a lot
of ways, it’s not a dramatic pivot. We’ve been doing distributed algebra and
analytics on Teradata forever – the SQL-based capabilities. Now you’re talking
about linear algebra, discrete math, calculus, differential equations. You’re
applying more sophisticated types of math. When you talk about deep learning,
you’re applying more sophisticated math on that data. But what everyone
struggles with is how you do that at scale. We’ve got the scaling part figured
out. You need to reach beyond just algebra into geometry, which is what you
need – Euclidean geometry in a lot of computer vision problems. But at the end
of the day, it’s just math at scale on data, and so that’s what we’re talking
about.
insideBIGDATA: And that’s what NVIDIA brings to the
table, yes? How are you guys working with NVIDIA?
Atif Kureishy: Absolutely.
We’ve been a
partner with NVIDIA for about one year, part of the Services Delivery Program (SDP).
If we engage with the customers and help them solve deep learning problems,
that’s going to push computation on the GPUs. So obviously, that’s very
harmonious.
Coming up next
year, we’re actually putting compute into our Vantage platform. You’re running
workload on the Teradata Vantage platform, and that data and computation will
be processed on GPUs for training, and serving the inference side. Ultimately,
you’re solving answers and problems for our customers. Our 2019 focus is
Vantage. We have all the computation and data, along with Teradata Everywhere,
AWS, and Azure. But let’s forget about all of that. The idea is if you can
deliver this in an “as-a-service” manner which really means in a more
consumable way to align a business executive.
We can do it
in a much more innovative and creative way using machine and deep learning. But
we’re not going to bring all that complexity. We’re going to give you a
subscription or some straightforward consumption-based method offering
dashboards, data pipelines, ML frameworks, data labeling/annotation schemes, and
GPU infrastructure. Every enterprise leader in the business wants all of that
sophistication without all the complexity, so that’s increasingly what we’re
focused on.
insideBIGDATA: What’s the timeframe for these
solutions?
Atif Kureishy: It’s an evolution. You’ll see this
carrying along a multi-year strategy. A lot of folks are doing this in the
cloud, so we embrace those partners where it makes sense. But the Fortune 100, what
we call “megadata” customers because of data gravity, privacy, security, etc. You
have to allow them to get to the cloud and that’s a part of our Teradata Everywhere
strategy. You also have to allow them to do analytics at scale in that same
Teradata Everywhere environment. By the way, deep learning is just an evolution
of ML. ML is just an evolution of some of the other modeling and simulation
techniques that we’ve been using. So you have to take customers on that path.
It’s
available, or will be available, on AWS and Azure, and on managed cloud. So
those things are available now, so folks can come on board now, and then when the
deep learning capabilities come out, they’ll have access to that technology as
well. It’ll be part of a first class environment with Vantage. The idea is that
we’re going to take them on that journey, and be there for them when they need
it.
insideBIGDATA: Can you describe a particularly use
case?
Atif Kureishy: Yes, there were some creative
applications of deep learning at Danske Bank with a variety of transactions
involving issuing bank, and receiving bank. We decided to extend and add new
features around everything else we know about the transactions, such as IP
addresses, Mac addresses, and other derivative information. Then we observed
that these transactions occur over time, so we were actually looking at
sequences of transactions rather than individual transactions. A lot of machine
learning approaches today look at a transaction in isolation in order to do comparative
analysis and anomaly detection. But we were actually looking at sequences of transactions
so there’s better signal in that detection.
So we took
the sequences arranged over time and we turn that into a model to emulate pixels
on an image. We literally took those transactional features and then did some
spatial correlations model techniques and we turned it into image.
Then we
applied convolutional neural networks (CCNs) to the image and that became a best-performing
method. We did time-aware LSTMs and other types of recurrent neural networks
(RNNs). The derivative benefit of this approach was that the auditors and regulators
could actually see fraud visually. We showed this kind of pixelization where
the intensity of a pixel would actually demonstrate fraud. They got it, and then
applied some other techniques to recognize attributes that contribute to the classifier
of false deny or approve. This was enough for us to understand what these black
box models are doing.
In the end,
this solution was an ensemble of six different techniques. We had some logistic
regression approaches, some boosted trees, and some other GLMs. Then we used a deep
neural network. It was such a dramatic improvement. We worked with them to
build their data science capabilities so that they could support this in the
future, and that’s why it was such a transformational effort.
insideBIGDATA: Well, this has been great. I appreciate
the opportunity to get a Teradata update.
Atif Kureishy: My pleasure.
Sign up for the free insideBIGDATA newsletter.
Leave a Reply