Note: The activities schedule on Wednesday are optional hands-on training for practitioners.
9:30 am - 11:30 am
1:00 pm - 3:00 pm
1:00 pm - 3:00 pm
3:30 pm - 5:30 pm
Training – Spark for Data Scientists
This presentation will start with a summary of Data Science and the importance by Piero Cinquegrana, Qubole's Data Science Product Manager. Subsequently Alex Aidun will review the features in Qubole that support Data Scientists during their development cycles - topics include Spark Notebooks, Qubole Features, Notebook API Execution, Notebook Dashboards, Notebook Interpreter Configuration.
9:00 am - 9:45 am
Opening Keynote – Big Data Activation
This keynote will discuss the gap that enterprises face today when activating their big data. It will make a case for the shift that organizations need to make towards a big data activation strategy in order to put their data assets to use for differentiating and achieving business objectives. The session will also cover key elements of big data activation supported by usage trends of Qubole's cloud-native big data activation platform. It will present various ways that enterprises can use to measure their own activation readiness, and demonstrate why Qubole provides the right approach to big data activation.
Ashish Thusoo,
Co-Founder & CEO,
Qubole
9:45 am - 10:15 am
Panel: Activating Big Data Across the Enterprise
Every CEO aspires to create a data-driven culture that can activate 100s or 1000s of users and petabyte-scale data to continuously deliver true business value. This keynote panel will explore the journey of 4 companies: Comcast, Qubole, Fanatics and MediaMath, that have chronicled their successes and challenges in two books by O’Reilly Media about Creating a Data-Driven Enterprises. The panelists will talk not just about their technology strategy and choices but also how data-driven insights are powering their business and transforming the competitive dynamics of their industry.
Santanu Dey,
Director of Data Science & Engineering,
Fanatics
Barbara Eckman,
Principal Data Architect,
Comcast
John Slocum,
Vice President,
Data Management Platform, MediaMath
Jose Villacis,
Sr. Director, Product Marketing,
Qubole
10:30 am - 11:15 am
Session I
Tech Talk – The Story of Building a Scalable Data Trust Playbook at Optimizely (Optimzely)
Key points:
At Optimizely, we receive billions of user click stream events for the thousands of A/B Experiments we run for our customers every day.
Customer inquiries related to alignment of key experiment metrics between raw data and what they see on their Experiment Results pages required expensive engineering analysis for lack of scale and flexibility in analysis.
In this talk, I will walk the audience through the journey of how we created a Playbook that can be used to enhance Customer Trust on Optimizely in these situations and made it scalable for the entire organization to self-serve.
Speaker: Vignesh Sukumar,
Senior Manager, Data Engineering,
Optimizely
10:30 am - 11:15 am
Session II
Tech Talk – A Lap Around Azure Data Lake (Insight)
Azure Data Lake is one of the most powerful PaaS services that Microsoft Azure offers to manage big data. Built on well-known projects such as HDFS and YARN, it allows for the ability to focus on the design of the solution instead of the administration part. A new language, U-SQL, combines SQL and C# to work with any type and size of data. During the session, we will explore Azure Data Lake Store and Azure Data Lake Analytics, the core components of the Azure Data Lake offering.
Speaker: Francesco Diaz,
Regional Solutions Manager Alps, Nordic & Southern Europe,
Insight
10:30 am - 11:15 am
Session I
Tech Talk – The BIG Picture: The Journey from On Premise Cloud to ML Marketplace for a Geospatial Data Platform (DigitalGlobe)
In a world driven by location intelligence, DigitalGlobe is creating a place where everyone can access geospatial data and use it to derive truth and, in turn, knowledge. DigitalGlobe’s Geospatial Big Data Platform (GBDX) is empowering an ecosystem of location intelligence, cataloging hundreds of petabytes worth of geospatial information and executing tens of millions (!) of hours’ worth of cloud compute.
This session will talk about the migration of GBDX from on-premise to cloud, involving 100 petabytes of satellite image data. We will also discuss GBDX as a key platform for analytics and machine learning applications across industries, and how it’s evolving into a marketplace for imagery-related machine learning algorithms.
Speaker: Ori Elkin,
Chief Product Officer,
Digital Globe
10:30 am - 11:15 am
Session I
10:30 am - 11:15 am
Session I
Tech Talk – Building a Real-Time Decision Engine Using Machine Learning on Apache Spark Structured Streaming (Blueprint Technologies)
Real-time decision making using ML/AI is the holy grail of customer-facing applications. It’s no longer a long-shot dream; it’s our new reality.
The real-time decision engine leverages the latest features in Apache Spark 2.3, including stream-to-stream joins and Spark ML, to directly improve the customer experience. We will discuss the architecture at length, including data source features and technical intricacies, as well as model training and serving dynamics. Critically, real-time decision engines that directly affect customer experience require production-level SLAs and/or reliable fallbacks to avoid meltdowns, which this talk will also address.
Speaker: Garren Staubli,
Sr. Data Engineer,
Blueprint Technologies
11:30 am - 12:15 pm
Session II
Tech Talk – A Framework for Assessing the Quality of Product Usage Data (Autodesk)
This presentation will discuss the importance of data quality and outline an approach to assess and measure the quality of product usage event logs. A data quality assessment framework helps build trust in our data and enables analysts to generate a deep understanding of product usage patterns, product stability, and utilization of purchased assets by our customers. The value that can be unlocked from this data is dependent upon high quality and complete data sets that provide the ability to link product usage events with back office account and entitlement data to provide valuable insights.
Speaker: David Oh,
Data Engineer for the ADP (Autodesk Data Platform),
Autodesk
11:30 am - 12:15 pm
Session II
Tech Talk – Presto: Fast SQL on Everything (Facebook)
Presto is an open source distributed query engine that supports much of the SQL analytics workload at Facebook. This talk introduces a selection of Facebook use cases, which range from user-facing reporting applications to multi-hour ETL jobs, then explains the architecture, implementation, features and performance optimizations that enable Presto to support these use cases.
Speaker: David Phillips,
Software Engineer,
Facebook
11:30 am - 12:15 pm
Session II
Tech Talk – Packaging, Deploying and Running Apache Spark Applications in Production (Mapbox)
Apache Spark has proven itself to be indispensable due to its endless applications and use cases. Developers, data scientists, engineers and analysts alike can benefit from its power. However, deterministically managing dependencies, packaging, testing, scheduling and deploying a Spark application can be challenging.
As organizations grow, these individuals become dispersed across multiple teams and departments. This makes a team specific solution no longer applicable. So, what type of tooling do you need to allow these individuals to solely focus on writing a Spark application? And more importantly, how do you enforce development best practices, such as unit testing, continuous integration, version control and deployment environments?
The data engineering team at Mapbox has developed tooling and infrastructure to address these challenges and enable individuals across the organization to build and deploy Spark applications. This talk will walk you through our solution to packaging, deploying, scheduling and running Spark applications in production. It will also address some of the problems we’ve faced and how the adoption process is evolving across the team.
Speaker: Saba El-Hilo,
Data Engineer,
Mapbox
11:30 am - 12:15 pm
Session II
Tech Talk – Lighthouse Related Product, an Efficient Cross-boundary Product Recommendation Platform on Qubole Ecosystem (Fanatics)
Fanatics, Inc. will introduce an item-to-item recommendation service platform in production, Lighthouse Related Product (LRP). LRP offers supervised-versus-non-supervised boundary-crossing and is extendable, flexible, and lightweight. LRP implements a modeling architecture that fuses into one system heterogeneous features from modern machine learning techniques of (1) non-supervised user-item matrices, (2) self-supervised Word2Vec, and (3) supervised XGBoost or deep learning. This architecture allows innate extendibility to user-item recommendations, and flexibility for both offline and online use cases. It is lightweight and efficient enough to handle near one million products' item-to-item recommendations on over 400 affiliated sites. The platform relies on the Apache Spark cluster in Qubole for both data feature extraction and prediction in a distributed manner with map procedure from a pre-trained supervised model; tasks on the Spark cluster in Qubole are seamlessly integrated into the rest of the workflows in Fanatics with another third-party scheduling service, Stone Branch. LRP has successfully passed the real-life load test of the 2017 holiday season and Super Bowl LII, and an earlier predecessor of the current version of LRP had achieved better performance in all measures, such as click-through rate and average order volume, compared to an industrial standard third-party recommendation service provider.
Speaker: Jing Pan,
User Experience Researcher,
Fanatics
11:30 am - 12:15 pm
Session II
Tech Talk – From Zero to Activated Big Data in the Cloud – the First Year’s Journey (Auris Surgical Robotics)
What would you do in this scenario?: a blank slate; one year to prepare for “big data is on the way”; and your company’s acknowledgement that data is a strategic corporate asset?
Attend this session to explore the goals, best practices, architectural constraints, and technologies that shaped the journey from that starting point to continuously delivered and live systems with engaged users. Also hear about the multiple use cases downstream from the data lake, such as APIs, guided exploration, streaming, and integration with external systems. Finally, learn how all of this can be accomplished by one full time employee and strategic partnerships.
Speaker: Brian Greene,
Cloud Data Architect,
Auris Surgical Robotics
1:30 pm - 2:15 pm
IBM Keynote: Demystifying AI, Machine Learning & Deep Learning
From chat bots, to recommendation engines, to Google Voice and Apple Siri, AI has begun to permeate our lives. We will demystify what AI is, present the difference between machine learning and deep learning, why the huge interest now, show some fun use cases and demos, and then discuss use cases of how deep learning based AI methods can be used to garner insights from data for enterprises. We will also talk about what IBM is doing to make deep learning and machine learning more accessible and useful to a broader set of data scientists.
Sumit Gupta,
VP, AI, Machine Learning and HPC,
IBM Cognitive Systems
2:30 pm - 3:15 pm
Session III
Tech Talk – Self Regulating Streaming Capabilities in Apache Heron (Streamlio)
Several enterprises have been producing data not only at high volume but also at high velocity. Many daily business operations depend on real-time insights, and therefore real-time staging and processing of the data is gaining significance. Thus, there is a need for a scalable infrastructure that can continuously ingest and process billions of events per day the moment the data is acquired.
To achieve real-time performance at scale, Twitter designed Heron for stream data processing. In production for more than four years, Heron faced crucial challenges from an operational point of view: the manual, time-consuming, and error-prone tasks of tuning various configuration knobs to achieve service level objectives (SLO), as well as the maintenance of SLOs in the face of sudden, unpredictable load variation and hardware or software performance degradation.
In order to address these issues, we conceived and implemented several innovative methods and algorithms that aim to bring self-regulating capabilities to these systems, thereby reducing the number of manual interventions. In this talk, we will give a brief introduction to issues and enumerate challenges such as slow hosts, unpredictable spikes, network slowness, and network partitioning that we faced in production. We'll also describe how we made the systems self-regulating to minimize overhead and operations.
Speaker: Karthik Ramasamy,
Co-Founder,
Streamlio / Twitter
2:30 pm - 3:15 pm
Session III
Tech Talk – Highly Scalable and Flexible ETL Tool Built on Top of Cascading Framework (BloomReach)
At BloomReach we have 100+ e-commerce customers sharing product catalogs that range from a few megabytes to hundreds of gigabytes, which then need to be parsed and transformed. In this presentation we will talk about how we built a custom ETL transformation tool on top of a cascading framework that handles custom transformations and joins at scale and speed.
Speaker: Navin Agarwal,
Principal Engineer,
BloomReach
2:30 pm - 3:15 pm
Session III
Tech Talk – Building Data Functions at Poshmark: From KPI Monitoring to Enabling Social Graphs (Poshmark)
Poshmark is the largest social marketplace for fashion in the U.S. where anyone can buy, sell, and share their personal style. With users engaging in 300M+ activities every day, data is a core asset at Poshmark. We began our data journey with very basic uses of data - monitoring high-level business KPIs. Four years later, we are now deploying data applications for actions such as enabling a balanced social graph and driving our homepage content based on real-time community activity. Join me in this session to hear key highlights from this incredible journey of building a data function at Poshmark, along with insights from the development of a people-matching algorithm and real-time user-driven homepage content.
Speaker: Barkha Saxena,
Vice President, Data & Analytics,
Poshmark
2:30 pm - 3:15 pm
Session III
Tech Talk – Using Qubole as the Data Lake for Programmatic Advertising (Adobe Ad Cloud)
Qubole has been the data warehouse of the DSP for the last six-plus years, and was selected as the ideal partner for mobilizing the considerable amount of diagnostic and base truth data contained within Amazon S3. From these origins, Qubole now powers our custom reporting infrastructure, machine learning algorithms, and user mapping reports, along with its evolving role in supporting system diagnoses and audits. We will touch on several use cases that demonstrate the flexibility and power of Qubole in democratizing data across the organization.
Speaker: Tom Silverstrim,
Sr. Manager, Adobe Media Optimizer,
Adobe Ad Cloud
2:30 pm - 3:15 pm
Session III
Tech Talk – Key Objectives and Principles for Building Predictive Models on Big Data (Neustar)
Business analysts spend a lot of time today looking at what happened in the past, but what about trying to grasp what will happen in the future? For example, what if you are given 10 percent more budget for next quarter’s marketing spend? Do you know how you’ll use that extra money, and do you know what impact it will create? Or suppose you want to increase your budget, but need to show what you expect that increase to do - then what?
Many of today’s data applications are simply “decision support systems” designed to be useful in the aforementioned scenarios. They help business professionals use data to better understand their environment and make better decisions. But with larger volumes of data and increased ambitions of competitive businesses, the end goals become tougher to achieve. As the VP of Engineering for MarketShare DecisionCloud at Neustar, which provides planning and analytics capabilities for marketers, Satya Ramachandran has taken on these challenges by leveraging big data technologies.
In this talk, Satya will discuss some of the high expectations he’s faced at MarketShare, and also some of his successes. For example, despite the fact that data has grown significantly in recent years, business users still want faster results. This phenomenon led to efforts that supported larger amounts of data within his organization and demanded speed improvements - going from several minutes to sub-second responses. Satya will share some guiding principles that helped him successfully develop and deploy the systems his customers needed to be successful with their big data projects.
Speaker: Satya Ramachandran,
Vice President, Engineering,
Neustar
3:30 pm - 4:15 pm
Session IV
Tech Talk – Optimize for Reduced Big Data Partitioning Costs (inMarket media, LLC)
Businesses that collect and process data can benefit greatly from partitioning their tables. Partitioning improves performance, increases query performance, and reduces the effort of rebuilding tables. Single partition queries can also be used to reduce the query load and avoid scanning the entire table.
However, transitioning large existing tables into partitioned tables can be cost-prohibitive. For example, we at inMarket media load billions of location records into multiple tables in our database to process these records through a pipeline of transformations. The resulting tables were not originally partitioned, and as time went on the decision not to partition the tables became increasingly expensive to maintain. We decided to partition our large existing tables to improve performance and reduce the costs of our queries. Initially, we thought that partitioning at this stage might be expensive. In my talk, I will give you a brief overview of the partitioning feature and explain the advantages and drawbacks of several different implementations of the partitioning process, as well as show how we were able reduce the cost of partitioning.
Speaker: Waad Aljaradt,
Data Engineer,
inMarket media LLC
3:30 pm - 4:15 pm
Session IV
Tech Talk – The 3S Method for Cluster Architecture Design (Oracle)
This session highlights the model used within Oracle Data Cloud (ODC) for Apache Hadoop 2 and Apache Spark clusters. We'll talk about taking the guesswork out of cluster design, and about the keys for balancing cost and performance while minimizing administrative overhead.
Speaker: Justin Wainwright,
Systems Analyst, Oracle Data Cloud,
Oracle
3:30 pm - 4:15 pm
Session IV
Tech Talk – Data as Reliable as Running Water: Analytics and Machine Learning (Uber)
Every aspect of the Uber experience is powered by data - everything from in-app ETAs, menu recommendations, and map labeling to driver dispatch and customer support. In this talk, we will focus on the infrastructure and platforms that power data ingestion, storage, streaming/batch analytics, and machine learning for thousands of operators, data scientists, and engineers at Uber.
Speaker: Nikhil Joshi,
Senior Product Manager, Data Infrastructure and Data Platforms,
Uber
3:30 pm - 4:15 pm
Session V
4:30 pm - 5:15 pm
Session V
Supercharging Performance of Spark Applications
Spark applications are difficult to tune for optimal performance and with the use of cloud stores like S3 as truth-store makes things even more complex. This talk will cover briefly about SparkLens (Spark tuning tool), Spark with Rubix (distributed cache), direct-write for Hive tables and its perf numbers.
Speaker: Venkat Sowrirajan,
Software Engineer,
Qubole
4:30 pm - 5:15 pm
Session V
Faster SQL on Hadoop on Cloud Platforms for Ad-hoc & Interactive Analysis
Popular SQL on Hadoop engines like SparkSQL on Apache Spark, Hive and Presto have got must faster on the cloud. This talk will explore the major features, architectural changes and best practices to super charge these SQL engines. The talk will also peek into the future for upcoming performance-related features.
Speaker: Rajat Venkatesh,
Sr. Dir. of Engineering,
Qubole
4:30 pm - 5:15 pm
Session V
Reimagining Data Platforms – Design Thinking and Innovation Workshop
Join us for a fun and interactive hands-on session to reimagine and design your ideal data platforms solution. Learn and use Design Thinking Methodology and add this Creative Next-Gen tool to your problem solving arsenal! Be a part of the exclusive club of innovators. Limited seating.
Speaker: Ankita Gautam,
User Experience Designer,
Qubole
4:30 pm - 5:15 pm
Using the Right Open Source Engine for the Right Job
Qubole has built a service that runs multiple engines on multiple clouds. In this talk, we will discuss the journey and the lessons learned, including insights from usage across multiple engines. The session will cover a range of topics such as running a reliable service, optimizing performance, increasing mean time to failure (MTTF) through improved monitoring, and implementing practices for staggered rollouts and fallbacks.
Namit Jain,
Senior Vice President, Engineering,
Qubole
4:30 pm - 5:15 pm
Auto Tuning Twitter Hadoop Jobs or: Don’t Touch That Analytics Dial!
Every day at Twitter, hundreds of thousands of Hadoop jobs transform and aggregate petabytes of data in our analytics stack. Historically, we've asked users to guess at tuning parameter values that affect how their Hadoop jobs run. For example, mapper and reducer counts, memory allocation, and intermediate serialization formats, among others. However, after looking at the values that users chose for tuning parameters in 2017, the data revealed that Hadoop jobs across our clusters were running sub-optimally and still not meeting users' Service Level Agreement (SLA) targets.
To address this problem, we implemented a service to automate the tuning of several of the most important Hadoop parameters, using historical per-job metrics to inform future runs. In this talk, we will review how the system works, some of the auto-tuning we've implemented so far, and what we have on our roadmap for the future.
Ben Pence,
Software Engineer,
Twitter
Anton Panasenko,
Software Engineer,
Twitter
6:30 pm - 9:30 pm
Dinner & Evening Event
You've spent the day learning how the wild west of big data is being won, now don your ten gallon hat and join us for a farm fresh dinner and ice cold saloon drinks. Watch the sunset on the patio of an beautiful adobe building, and network with other leaders and innovators in the big data space. This is a party not to be missed.
9:00 am - 9:30 am
9:45 am - 10:30 am
Session VI
Tech Talk – Kubernetes for Data Engineers (Google)
The talk will give an introduction to Kubernetes in general and then focus on topics relevant to Data Engineers: in particular, we will talk about how to run stateful workloads on Kubernetes and how to run Machine Learning workloads that use GPUs on Kubernetes.
Speaker: Rohit Agarwal,
Software Engineer,
Google
9:45 am - 10:30 am
Session VI
Tech Talk – Where We’re Going We Don’t Need Computers: End-to-End Serverless Data Science (Oracle)
Data scientists are expected to wear many hats in an organization. Many tasks often fall in the realm of data science - ingesting and cleaning data, managing data storage, creating scalable machine learning models, and publishing APIs to expose and schedule services for end users. This talk focuses on how to create end-to-end data science products that allow data scientists to focus on business logic, all while embracing nearly infinitely horizontally scalable data platforms.
To do this, we’ll explore serverless cloud technologies at multiple levels of the data science pipeline such as serverless compute, workflow, containerized workloads, distributed on-demand machine learning, metrics tracking, and API as a service. At the end of this talk we'll have a prototype for an end-to-end machine learning system, on a scalable cloud platform, capable of processing petabytes of data and thousands of requests without the need for any freestanding servers.
Speaker: Alex Sadovsky,
Senior Director of Data Science,
Oracle Data & Cloud
9:45 am - 10:30 am
Session VI
Enterprise Fabric – A Concept for the Essential Thread in Your Transformational Journeys
This session will provide an overview of the enterprise fabric and the encapsulated view of the required capabilities. Some key components of the fabric include data and cognitive technologies. We will dive into the enterprise fabric-based architecture and why it is the core foundation for business transformation.
Speaker: Dan Sutherland,
Distinguished Engineer & CTO - Data Platforms, Global Business Services,
IBM
9:45 am - 10:30 am
Session VI
Tech Talk – Big Data + Data Warehouse = Better Together (Microsoft)
Building a coherent platform for advanced analytics and reporting can feel quite overwhelming. A plethora of choices exist out there, and often it can feel like you are being asked to choose a side - it's almost like being asked to pick your favorite child! However, it doesn't have to be this way. Many of the services in a next-generation platform are actually complimentary, working together to deliver your next-generation analytical architecture.
In this session we will walk through the core components of a next-generation analytical platform architecture, discussing key decision points along the way. At the end you will have a clear and concrete understanding of how you can easily stand up an advanced analytical platform in minutes and bring demonstrable value to your users.
Speaker: James Rowland-Jones,
Principal Program Manager, Engineering,
Microsoft
9:45 am - 10:30 am
Session VI
Tech Talk – Building Your Data Lake on Amazon S3: Architecture and Best Practices (AWS)
Learn how to build and architect a data lake on AWS where different teams within your organization can publish and consume data in a self-service manner. And learn about best practices for data curation, normalization, and analysis on Amazon object storage services. As organizations aim to become more data-driven, data engineering teams have to build architectures that can cater to the needs of diverse users - from developers, to business analysts, to data scientists. Each of these user groups employs different tools, have different data needs and access data in different ways.
If we end up doing the Quickstart demo, then I can just include that in the session.
Speaker: Paul Sears,
Solutions Architect, AWS Partner Network,
Amazon Web Services
10:45 am - 11:30 am
Session VII
Tech Talk – The Dismal and Uncomfortable Science of Data Engineering: Building Out Big Data with Your Analytics Team (Jumis)
While software services catch up, many Big Data projects rely on engineering resources, people with programming skills and the culture around software development. Meanwhile, analysis teams are often oriented to serve customer engagement and finance managers, with an ethos and engagement style quite distinct from their counterparts in the school of computer science. As business and engineering sides of the house clash and cooperate, it's important to remember the human side of things, both in the data and in the delivery of insights. Let's talk about ways these capable Business Analysts and Data Engineers can work together, and ideals they can align toward.
Speaker: Charles Pritchard,
Data Janitor,
Jumis
10:45 am - 11:30 am
Session VII
Tech Talk – Velocity Versus Volume (Expedia)
For technology companies, there is an inherent tension between streaming and batch processing. Real-time datastreams can transform a small input signal into an immediate response, but machine learning is most effective in batch. Modern data platforms can easily handle both streaming and batch jobs simultaneously. Balancing these two paradigms thus becomes a matter of design, and right now this interplay is thriving at the intersection of product and data science.
We discuss these dualities in the context of recommendations systems, some of our core products at Expedia. We’ll sketch the design, architecture, tools, and metrics, as well as share our experience with our attempts at personalization. We’ll merge the ideas behind multi-armed bandits and learning-to-rank to develop a novel recommendation system and give you the background needed to start building products in this rapidly evolving space.
Speaker: Sean Downes,
Data Scientist,
Expedia
10:45 am - 11:30 am
Session VII
Tech Talk – Team Data Science Process (TDSP) and Azure Machine Learning (Microsoft)
TDSP is an agile data science process meant to keep data science and business teams working together. In this session, we'll explore the Team Data Science Process and walk through an example using Azure Machine Learning Services.
Speaker: Erik Zwiefel,
Advanced Analytics and AI Architect,
Microsoft
10:45 am - 11:30 am
Session VII
10:45 am - 11:30 am
Session VII
Tech Talk – Email Text Classification: Building an End-to-End Data Product (Return Path)
Sasha will tell the story of building an end-to-end data product that feeds various parts of the Return Path business to optimize email programs for marketers. We will cover discovery, development, and production of an email classification model that uses Apache Spark to fit classifiers such as Random Forests and Support Vector Machines to read email text and classify the content. We will discuss the different methods of hyperparameter tuning and ensembling used, and will describe different stages of production from batch jobs in Qubole Scheduler and Apache Airflow to streaming in Apache Kafka. We will also reflect on what it means to be a full stack data scientist, and how data science teams can be empowered to own their own data products.
Speaker: Sasha Mushovic,
Data Scientist,
Return Path