For audio versions: Podcast RSS Feed
Podcast version also available on iTunes
Andy Jassy, CEO Amazon Web Services
CEO Andy Jassy discusses the latest AWS news and announcements.
Werner Vogels, CTO Amazon Web Services
CTO Werner Vogels discusses the latest AWS news and announcements.
James Hamilton, Distinguished Engineer, Amazon Web Services
Distinguished Engineer James Hamilton
David Isbitski - Chief Evangelist, Alexa & David Isbitski - Chief Evangelist, Alexa and Echo & Scott Totman - VP, Mobile Engineering, Capital One
As we add thousands of skills to Alexa, our developers have uncovered some basic and more complex tips for building better skills. Whether you are new to Alexa skill development or if you have created skills that are live today, this session helps you understand how to create better voice experiences. Last year, Capital One joined Alexa on stage at re:Invent to talk about their experience building an Alexa skill. Hear from them one year later to learn from the challenges that they had to overcome and the results they are seeing from their skill. In this session, you will learn the importance of flexible invocations, better VUI design, how OAuth and account linking can add value to your skill, and about Capital One's experience building an Alexa skill.
CJ Frost - Principal Solutions Architect - Alexa Automotive, Amazon
The experience in the auto industry is changing. For both the driver and the car manufacturer, a whole new frontier is on the near horizon. What do you do with your time while the car is driving itself? How do I have a consistent experience while driving shared or borrowed cars? How do I stay safer and more aware in the ever increasing complexity of traffic, schedules, calls, messages and tweets? In this session we will discuss how the auto industry is facing new challenges and how the use of Amazon Alexa, IoT, Logistics services and the AWS Cloud is transforming the Mobility experience of the (very near) future.
Jeff Blankenburg & Jeff Blankenburg - Sr. Alexa Evangelist
This workshop teaches you how to build your first voice skill with Alexa. You bring a skill idea and we’ll show you how to bring it to life. This workshop will walk you through how to build an Alexa skill, including Node.js setup, how to implement an intent, deploying to AWS Lambda, and how to register and test a skill. You’ll walk out of the workshop with a working prototype of your skill idea.
Amit Jotwani - Senior Evangelist & Amit Jotwani
Fascinated by Alexa, and want to build your own device with Alexa built in? This workshop will walk you through to how to build your first Alexa-powered device step by step, using a Raspberry Pi. No experience with Raspberry Pi or Alexa Voice Service is required. We will provide you with the hardware and the software required to build this project, and at the end of the workshop, you will be able to walk out with a working prototype of Alexa on a Pi.
Jeff Blankenburg & Jeff Blankenburg - Sr. Alexa Evangelist & Michael (Mik) Cox - Data Scientist, Jet Propulsion Laboratory & Tom Soderstrom - IT Chief Technology and Innovation Officer, JPL
The Jet Propulsion Laboratory designs and creates some of the most advanced space robotics ever imagined. JPL IT is now innovating to help streamline how JPLers will work in the future in order to design, build, operate, and support these spacecraft. They hope to dramatically improve JPLers' workflows and make their work easier for them by enabling simple voice conversations with the room and the equipment across the entire enterprise.
Sebastien Stormacq - Solution Architect
Learn how to develop voice-based serverless back ends for Alexa Voice Service (AVS) and Alexa devices using the Alexa Skills Kit (ASK), which allows you to add new voice-based interactions to Alexa. We’ll code a new skill, implemented by a serverless backend leveraging AWS services such as Amazon Cognito, AWS Lambda, and Amazon DynamoDB. Often, your skill needs to authenticate your users and link them back to your backend systems and to persist state between user invocations. User authentication is performed by leveraging OAuth compatible identity systems. Running such a system on your back end requires undifferentiated heavy lifting or boilerplate code. We’ll leverage Login with Amazon as the identity provider instead, allowing you to focus on your application implementation and not on the low-level user management parts. At the end of this session, you’ll be able to develop your own Alexa skills and use Amazon and AWS services to minimize the required backend infrastructure. This session shows you how to deploy your Alexa skill code on a serverless infrastructure, leverage AWS Lambda, use Amazon Cognito and Login with Amazon to authenticate users, and leverage AWS DynamoDB as a fully managed NoSQL data store.
Natural user interfaces, such as those based on speech, enable customers to interact with their home in a more intuitive way. With the VUI (Voice User Interface) smart home, now customers don't need to use their hands or eyes to do things around the home — they only have to ask and it's at their command. This session will address the vision for the VUI smart home and how innovations with Amazon Alexa make it possible.
Matt Tavis - Principal Solutions Architect & Matt Tavis - Principal Solutions Architect
Ever wonder what it takes to add the power of Alexa to your own products? Are you curious about what Alexa partners have learned on their way to a successful product launch? In this session you will learn about the top tips and tricks on how to go from VUI newbie to an Alexa-enabled product launch. Key concepts around hardware selection, enabling far field voice interaction, building a robust Alexa Voice Service (AVS) client and more will be discussed along with customer and partner examples on how to plan for and avoid common challenges in product design, development and delivery.
Paul Cutsinger , Amazon Web Services & Andrew Huntwork - Principal Engineer & Andy Huntwork
Hitting the submit button to publish your skill is similar to sending your child to their first day of school. You want it to be set up for a successful launch day and for many days thereafter. Learn how to set your skill up for success from Andy Huntwork, Alexa Principal Engineer and one of the creators of the popular Alexa skill 'The Magic Door'. You will learn the most common reasons why skills fail and also some of the more unique use cases. The purpose of this session is to help you build better skills by knowing what to look out for and what you can test for before submitting. In this session, you will learn what most developers do wrong, how to successfully test and QA your skill, how to set your skill up for successful certification, and the process of how a skill gets certified.
Rohit Prasad - VP and Head Scientist, Alexa
The way humans interact with machines is at a turning point, and conversational artificial intelligence (AI) is at the center of the transformation. Learn how Amazon is using machine learning and cloud computing to fuel innovation in AI, making Amazon Alexa smarter every day. Alexa VP and Head Scientist Rohit Prasad presents the state of the union Alexa and Recent Advances in Conversational AIn for Alexa. He addresses Alexa's advances in spoken language understanding and machine learning, and shares Amazon's thoughts about building the next generation of user experiences.
Nathan Grice - Senior Solutions Architect - Alexa & nathan grice & Tom Wynnyk & Mark Gregas - Director of Information Technology, Brookfield Residential Properties
Want to learn how to Alexa-power your home? Join Brookfield Residential CIO and EVP Tom Wynnyk and Senior Solutions Architect Nathan Grice, for Alexa Smart Home for an overview of building the next generation of integrated smart homes using Alexa to create voice-first experiences. Understand the technologies used and how to best expose voice experiences to users through Alexa. Paul and Nathan cover the difference between custom Alexa skills and Smart Home Skill API skills, and build a home automation control from the ground up using Alexa and AWS IoT.
Joel Williams - AWS Solutions Architect, AWS
Cloud computing gives you a number of advantages, such as the ability to scale your web application or website on demand. If you have a new web application and want to use cloud computing, you might be asking yourself, 'Where do I start?' Join us in this session to understand best practices for scaling your resources from zero to millions of users. We show you how to best combine different AWS services, how to make smarter decisions for architecting your application, and how to scale your infrastructure in the cloud.
Thomas Myers - Senior Cloud Architect, Accenture & Matthew Lancaster - Principal Director, Accenture
Accenture Cloud Platform helps customers manage public and private enterprise cloud resources effectively and securely. In this session, learn how we designed and built new core platform capabilities using a serverless, microservices-based architecture that is based on AWS services such as AWS Lambda and Amazon API Gateway. During our journey, we discovered a number of key benefits, including a dramatic increase in developer velocity, a reduction (to almost zero) of reliance on other teams, reduced costs, greater resilience, and scalability. We describe the (wild) successes we’ve had and the challenges we’ve overcome to create an AWS serverless architecture at scale. Session sponsored by Accenture.
TIMOTHY DILAURO - Solutions Architect, Amazon Web Services & Joe Gardner - Principal Cloud Architect, National Instruments
The AWS Well-Architected Framework enables customers to understand best practices around security, reliability, performance, and cost optimization when building systems on AWS. This approach helps customers make informed decisions and weigh the pros and cons of application design patterns for the cloud. In this session, you'll learn how National Instruments used the Well-Architected Framework to follow AWS guidelines and best practices. By developing a strategy based on the AWS Well-Architected Framework, National Instruments was able to triple the number of applications running in the cloud without additional head count, significantly increase the frequency of code deployments, and reduce deployment times from two weeks to a single day. As a result, National Instruments was able to deliver a more scalable, dynamic, and resilient LabVIEW platform with agility.
Coburn Watson - Director, Performance and Reliability, Netflix, Inc
Building and evolving a pervasive, global service requires a multi-disciplined approach that balances requirements with service availability, latency, data replication, compute capacity, and efficiency. In this session, we’ll follow the Netflix journey of failure, innovation, and ubiquity. We'll review the many facets of globalization and then delve deep into the architectural patterns that enable seamless, multi-region traffic management; reliable, fast data propagation; and efficient service infrastructure. The patterns presented will be broadly applicable to internet services with global aspirations.
Paul Underwood - Solution Architect, Amazon Web Services & Keith Horwood - Founder and CEO, Polybit Inc.
This presentation provides a comparison of three modern architecture patterns that startups are building their business around. It includes a realistic analysis of cost, team management, and security implications of each approach. It covers Elastic Beanstalk, Amazon ECS, Docker, Amazon API Gateway, AWS Lambda, Amazon DynamoDB, and Amazon CloudFront, as well as Docker.
Ramanuja Atur - Senior Product Manager - Technical, Amazon Web Services & Randy Young - Principal Product Manager, Splunk
When your business is counting on the performance of your cloud solutions, having relevant and timely insights into events impacting your AWS resources is essential. AWS Personal Health Dashboard serves as the primary destination for you to receive personalized information related to your AWS infrastructure, guiding your through scheduled changes, and accelerating the troubleshooting of issues impacting your AWS resources. The service, powered by AWS Health APIs, integrates with your in-house event management systems, and can be programmatically configured to proactively get the right information into the right hands at the right time. The service is integrated with Splunk App for AWS to enhance Splunk’s dashboards, reports and alerts to deliver real-time visibility into your environment.
Jamie Butler - Manager, Solutions Architecture, Amazon Web Services
AWS provides many services to assist customers with their journey to the cloud. Hybrid solutions offer customers a way to continue leveraging existing investments on-premises, while expanding their footprint into the public cloud. This session covers the different technologies available to support hybrid architectures on AWS. We discuss common patterns and anti-patterns for solving enterprise workloads across a hybrid environment.
Eric Johnson - AWS Evangelist, Rackspace & James Cowe - Director, AWS Strategy and Architecture, Rackspace
In today’s world, technology changes at a breakneck speed. What was new this morning is outdated at lunch. Working in the AWS Cloud is no different. Every week, AWS announces new features or improvements to current products. As AWS technologists, we must assimilate these new technologies and make decisions to adopt, reject, or defer. These decisions can be overwhelming: we tend to either reject everything and become stagnant, or adopt everything and never get our project out the door. In this session we will discuss the attitude of iteration. The attitude of iteration allows us to face the challenges of change without overwhelming our technical teams with a constant tug-o-war between implementation and improvement. Whether you’re an architect, engineer, developer, or AWS newbie, prepare to laugh, cry, and commiserate as we talk about overcoming these challenges.
Kevin Volk - Sr. Program Manager, Amazon Web Services
Come and participate with other AWS customers as we focus on the overall experience of using AWS to solve business problems. This is a great opportunity to collaborate with existing and prospective AWS users to validate your thinking and direction with AWS peers, discuss the resources that aid AWS solution design, and give direct feedback on your experience building solutions on AWS.
Sean Senior - Solutions Architect, Amazon Web Services
Regularly, customers at AWS assign resources to create solutions that address common problems shared between businesses of all sizes. Often, this results in taking resources away from products or services that truly differentiate the business in the marketplace. The Solutions Builder team at AWS focuses on developing and publishing a catalog of repeatable, standardized solutions that can be rapidly deployed by customers to overcome common business challenges. In this session, the Solutions Builder team will share ready to use solutions that make it easy for anyone to create a transit VPC, centralized logging, a data lake, scheduling for Amazon EC2, and VPN monitoring. Along the way, the team reveals the architectural tenets and best practices they follow for the development of these solutions. In the end, customers are introduced to a catalog of freely available solutions with a peek into the architectural approaches used by an internal team at AWS.
Varun Gupta - Director of Engineering, Salesforce & Tyson Lutz - Senior Director of Internal Capacity Engineering, Salesforce & Vijaysenthil Veeriah - Software Engineer, Salesforce.com
Salesforce is one of the most innovative enterprise software companies in the world, delivering 3 major releases a year with hundreds of features in each release. In this session, come learn how we enable thousands of engineers within Salesforce to utilize a flexible development environment to deliver these innovations to our customers faster. We show you how we enable engineers at Salesforce to test not only individual services they are developing but also large scale service integrations. Also learn how we can achieve setup of a representative production environment in minutes and teardown in seconds, using AWS.
Adrian Cockcroft - VP Cloud Architecture Strategy, AWS
Over the last few years, we have seen a dramatic increase in the use of open source projects as the mainstay of architectures in both startups and enterprises. Many of our customers and partners also run their own open source programs and contribute key technologies to the industry as a whole (see DCS201). At AWS we engage with open source projects in a number of ways. We contribute bug fixes and enhancements to popular projects including our work with the Hadoop ecosystem (see BDM401), Chromium (see BAP305) and (obviously) Boto. We have our own standalone projects including the security library s2n (see NET405) and machine learning project MXnet (see MAC401). We also have services that make open source easier to use like ECS for Docker (see CON316), and RDS for MySQL and PostgreSQL (see DAT305). In this session you will learn about our existing open source work across AWS, and our next steps.
Tod Golding - Partner Solutions Architect, Amazon Web Services
AWS provides a broad array of services, tools, and constructs that can be used to design, operate, and deliver SaaS applications. In this session, Tod Golding, the AWS Partner Solutions Architect, shares the wisdom and lessons learned from working with dozens of customers and partners building SaaS solutions on AWS. We discuss key architectural strategies and patterns that are used to deliver multi-tenant SaaS models on AWS and dive into the full spectrum of SaaS design and architecture considerations, including tenant isolation models, tenant identity management, serverless SaaS, and multi-tenant storage strategies. This session connects the dots between general SaaS best practices and what it means to realize these patterns on AWS, weighing the architectural tradeoffs of each model and assessing its influence on the agility, manageability, and cost profile of your SaaS solution.
Rob Alexander - AWS Solutions Architect, Amazon Web Services
As more customers adopt Amazon VPC architectures, the features and flexibility of the service are squaring off against evolving design requirements. This session follows this evolution of a single regional VPC into a multi-VPC, multi-region design with diverse connectivity into on-premises systems and infrastructure. Along the way, we investigate creative customer solutions for scaling and securing outbound VPC traffic, securing private access to Amazon S3, managing multi-tenant VPCs, integrating existing customer networks through AWS Direct Connect, and building a full VPC mesh network across global regions.
Lee Atchison - Principal Cloud Architect, New Relic
Applications running in a typical data center are static entities. Dynamic scaling and resource allocation are the norm in AWS. Technologies such as Amazon EC2, Docker, AWS Lambda, and Auto Scaling make tracking resources and resource utilization a challenge. The days of static server monitoring are over.
Ben Abrams - Supreme Unicorn Hunter of Planet Earth and the Entire Galaxy Besides, Cloud Cruiser & Ben Newton - Principle Product Manager, Sumo Logic & Christian Beedgen - CTO & Co-Founder, Sumo Logic
IT is evolving from a cost center to a source of continuous innovation for business. At the heart of this transition are modern, revenue-generating applications, based on dynamic architectures that constantly evolve to keep pace with end-customer demands. This dynamic application environment requires a new, comprehensive approach to traditional monitoring – one based on real-time, end-to-end visibility and analytics across the entire application lifecycle and stack, instead of monitoring by piecemeal. This presentation highlights practical advice on how developers and operators can leverage data and analytics to glean critical information about their modern applications. In this session, we will cover the types of data important for today’s modern applications. We’ll discuss visibility and analytics into data sources such as AWS services (e.g., Amazon CloudWatch, AWS Lambda, VPC Flow Logs, Amazon EC2, Amazon S3, etc.), development tool chain, and custom metrics, and describe how to use analytics to understand business performance and behaviors. We discuss a comprehensive approach to monitoring, troubleshooting, and customer usage insights, provide examples of effective data analytics to improve software quality, and describe an end-to-end customer use case that highlights how analytics applies to the modern app lifecycle and stack.
Derek Chiles - Senior Manager, AWS Tech Leader Team, Amazon Web Services & Emerson Loureiro - Senior Software Engineer, Gilt Groupe
Gilt, a global e-commerce company, implemented a sophisticated microservices architecture on AWS to handle millions of customers visiting their site at noon every day. The microservices architecture pattern enables independent service scaling, faster deployments, better fault isolation, and graceful degradation. In this session, Emerson Loureiro, Sr. Software Engineer at Gilt, will share Gilt's experiences and lessons learned during their evolution from a single monolithic Rails application in a traditional data center to more than 300 Scala/Java microservices deployed in the cloud. Derek Chiles, AWS Solutions Architect, will review best practices and recommended architectures for deploying microservices on AWS.
Benjamin Feldon - Enterprise Solutions Architect, Amazon Web Services & Theresa Sage - VP, Engineering, McGraw-Hill Education
How does McGraw-Hill Education use the AWS platform to scale and reliably receive 10,000 learning events per second? How do we provide near-real-time reporting and event-driven analytics for hundreds of thousands of concurrent learners in a reliable, secure, and auditable manner that is cost effective? MHE designed and implemented a robust solution that integrates AWS API Gateway, AWS Lambda, Amazon Kinesis, Amazon S3, Amazon Elasticsearch Service, Amazon DynamoDB, HDFS, Amazon EMR, Amazopn EC2, and other technologies to deliver this cloud-native platform across the US and soon the world. This session describes the challenges we faced, architecture considerations, how we gained confidence for a successful production roll-out, and the behind-the-scenes lessons we learned.
Fred Rahmanian - Chief Technology Officer, Geneia & Raja Renganathan - Vice President, Cognizant Cloud Services, Cognizant
Hear Geneia's design principles for using multiple technologies like Elastic Load Balancing and Auto Scaling in end-to-end solutions to meet regulatory requirements. Explore how to meet HIPAA regulations by using native cloud services like Amazon EC2, Amazon EBS volumes, encryption services, and monitoring features in addition to third-party tools to ensure end-to-end data protection, privacy, and security for protected health information (PHI) data hosted in the AWS Cloud. Learn how Geneia leveraged multiregion and multizone backup and disaster recovery solutions to address the recovery time objective (RTO) and recovery point objective (RPO) requirements. Discover how automated build, deployment, provisioning, and virtual workstations in the cloud enabled Geneia's developers and data scientists to quickly provision resources and work from any location, expediting the onboarding of customers, getting to market faster, and capturing bigger market share in healthcare analytics while minimizing costs. Session sponsored by Cognizant.
Diego Macadar - Software Development Manager, Amazon Web Services
Learn how AWS processes millions of records per second to support accurate metering across AWS and our customers. This session shows how we migrated from traditional frameworks to AWS managed services to support a large processing pipeline. You will gain insights on how we used AWS services to build a reliable, scalable, and fast processing system using Amazon Kinesis, Amazon S3, and Amazon EMR. Along the way we dive deep into use cases that deal with scaling and accuracy constraints. Attend this session to see AWS’s end-to-end solution that supports metering at AWS.
David Green - Enterprise Solutions Architect, Amazon Web Services & Alexander Filipchik - Principal Software Engineer, Sony Interactive entertainment & Dustin Pham - Principal Engineer, Sony Interactive Entertainment
In gaming, low latencies and connectivity are bare minimum expectations users have while playing online on PlayStation Network. Alex and Dustin share key architectural patterns to provide low latency, multi-region services to global users. They discuss the testing methodologies and how to programmatically map out a large dependency multi-region deployment with data-driven techniques. The patterns shared show how to adapt to changing bottlenecks and sudden, several million request spikes. You’ll walk away with several key architectural patterns that can service users at global scale while being mindful of costs.
Simon Elisha - Head of Solution Architecture, Public Sector, Australia & New Zealand, Amazon Web Services
Did you know that AWS enables builders to architect solutions for price? Beyond the typical challenges of function, performance, and scale, you can make your application cost effective. Using different architectural patterns and AWS services in concert can dramatically reduce the cost of systems operation and per-transaction costs. This session uses practical examples aimed at architects and developers. Using code and AWS CloudFormation in concert with services such as Amazon EC2, Amazon ECS, Lambda, Amazon RDS, Amazon SQS, Amazon SNS, Amazon S3, CloudFront, and more, we demonstrate the financial advantages of different architectural decisions. Attendees will walk away with concrete examples, as well as a new perspective on how they can build systems economically and effectively.
Thomas Lockney - Senior Engineering Manager, Nike & Owyn Richen - Sr. Director Platform Delivery, Nike
Nike+ is at the core of the Nike digital product ecosystem, providing services to enhance your athletic experience through quantified activity tracking and gamification. As one of the first movers at Nike to migrate out of the datacenter to AWS, they share the evolution in building a reactive platform on AWS to handle large, complex data sets. They provide a deep technical view of how they process billions of metrics a day in their quantified-self platform, supporting millions of customers worldwide. You’ll leave with ideas and tools to help your organization scale in the cloud. Come learn from experts who have built an elastic platform using Java, Scala, and Akka, leveraging the power of many AWS technologies like Amazon EC2, ElastiCache, Amazon SQS, Amazon SNS, DynamoDB, Amazon ES, Lambda, Amazon S3, and a few others that helped them (and can help you) get there quickly.
Peter O'Donnell - Solutions Architect, Amazon Web Services & Kapil Thangavelu - Director of Engineering, Capital One
What happens when you give 6,000 developers access to the cloud? Introducing Cloud Custodian, an open source project from Capital One, which provides a DSL for AWS fleet management that operates in real-time using CloudWatch Events and Lambda. Cloud Custodian is used for the gamut of compliance, encryption, and cost optimization. What can it do for you?
Constantin Gonzalez - AWS Principal Solutions Architect, Amazon Web Services & Markus Ostertag - Organizer AWS Usergroup Munich, Head of Development Team Internet AG,
Whether you’re a cash-strapped startup or an enterprise optimizing spend, it pays to run cost-efficient architectures on AWS. This session reviews a wide range of cost planning, monitoring, and optimization strategies, featuring real-world experience from AWS customers. We cover how to effectively combine Amazon EC2 On-Demand, Reserved, and Spot instances to handle different use cases; leveraging Auto Scaling to match capacity to workload; choosing the optimal instance type through load testing; taking advantage of Multi-AZ support; and using Amazon CloudWatch to monitor usage and automatically shut off resources when they are not in use. We discuss taking advantage of tiered storage and caching, offloading content to Amazon CloudFront to reduce back-end load, and getting rid of your back end entirely by leveraging AWS high-level services. We also showcase simple tools to help track and manage costs, including Cost Explorer, billing alerts, and AWS Trusted Advisor. This session is your pocket guide for running cost effectively in the Amazon Cloud.
Koen Biggelaar - Senior Manager, AWS Solutions Architecture, Amazon Web Services & Henk van Rossum - Director Platform and Program Management Hosting & Storage, Philips
With customers migrating workloads to AWS, we are starting to see a need for the creation of a prescribed landing zone, which uses native AWS capabilities and meets or exceeds customers' security and compliance objectives. In this session, we will describe an AWS landing zone and will cover solutions for account structure, user configuration, provisioning, networking and operation automation. This solution is based on AWS native capabilities such as AWS Service Catalog, AWS Identity and Access Management, AWS Config Rules, AWS CloudTrail and Amazon Lambda. We will provide an overview of AWS Service Catalog and how it be used to provide self-service infrastructure to applications users, including various options for automation. After this session you will be able to configure an AWS landing zone for successful large scale application migrations. Additionally, Philips will explain their cloud journey and how they have applied their guiding principles when building their landing zone.
Constantin Gonzalez - AWS Principal Solutions Architect, Amazon Web Services & Christian Deger - Chief Architect, AutoScout24 GmbH
Fed up with stop and go in your data center? Shift into overdrive and pull into the fast lane!
Alan Halachmi - Sr. Manager, Solutions Architecture, Amazon Web Services
This session demonstrates how customers can leverage hybrid IT as a transitional step on the path to going all-in on AWS. We provide a step-by-step walk-through focusing on seamless migration to the cloud, with consideration given to existing data centers, equipment, and staff retraining. Learn about the suite of capabilities AWS provides to ease and simplify your journey to the cloud.
Benjamin Snively - Solutions Architect, Amazon Web Services & Stan Przychodzki - Manager, SaaS Architecture & Performance Engineering, D2L & Stephen Skrzydlo - Business Intelligence Architect, D2L Corporation
When D2L first moved to the cloud, we were concerned about being locked-in to one cloud provider. We were compelled to explore the opportunities of the cloud, so we overcame our perceived risk, and turned it into an opportunity by self-rolling tools and avoiding AWS native services. In this session, you learn how D2L tried to bypass the lock but eventually embraced it and opened the cage. Avoiding AWS native tooling and pure lifts of enterprise architecture caused a drastic inflation of costs. Learn how we shifted away from a self-rolled 'lift' into an efficient and effective 'shift' while prioritizing cost, client safety, AND speed of development. Learn from D2L's successes and missteps, and convert your own enterprise systems into the cloud both through native cloud births and enterprise conversions. This session discusses D2L’s use of Amazon EC2 (with a guest appearance by Reserved Instances), Elastic Load Balancing, Amazon EBS, Amazon DynamoDB, Amazon S3, AWS CloudFormation, AWS CloudTrail, Amazon CloudFront, AWS Marketplace, Amazon Route 53, AWS Elastic Beanstalk, and Amazon ElastiCache.
Patrick McClory - Senior Vice President, Platform Engineering and Delivery Services, Datapipe
For an IT organization to be successful in rapid cloud assessment or iterative migration of their infrastructure and applications to AWS, they need to effectively plan and execute on a strategic cloud strategy that focuses not only on cloud, but also big data, DevOps, and security.
ARC320: Workshop: AWS Professional Services Effective Architecting Workshop
Mark Statham - Senior Cloud Architect, Amazon Web Services
The AWS Professional Services team will be facilitating an architecture workshop exercise for certified AWS Architects. Class size will be limited to 48. This workshop will be a highly interactive architecture design exercise where the class will be randomly divided into teams and given a business case for which they will need to design an effective AWS solution. Past participants have found the interaction with people from other organizations and the creative brainstorming that occurs across 6 different teams greatly enhances the learning experience. Flipcharts will be provided and students are encouraged to bring their laptops to document their designs. Each team will be expected to present their solution to the class.
Drew Dennis - AWS Solutions Architect, Amazon Web Services & Maitreya Ranganath - AWS Solutions Architect, Amazon Web Services & Ajoy Kumar - Architect, BMC Software
As serverless architectures become more popular, AWS customers need a framework of patterns to help them deploy their workloads without managing servers or operating systems. This session introduces and describes four re-usable serverless patterns for web apps, stream processing, batch processing, and automation. For each, we provide a TCO analysis and comparison with its server-based counterpart. We also discuss the considerations and nuances associated with each pattern and have customers share similar experiences. The target audience is architects, system operators, and anyone looking for a better understanding of how serverless architectures can help them save money and improve their agility.
Dhruv Thukral - Solutions Architect, Amazon Web Services & William Day - Senior Software Engineer, WB Games: Turbine & Kyle Boorky - Staff Engineer, WB Games: Turbine
Warner Bros’ Turbine team shares lessons learned from their enhanced microservices game platform, which uses Docker, Amazon EC2, Elastic Load Balancing, and Amazon ElastiCache to scale up in anticipation of massive game adoption. Learn about their Docker-based microservices architecture, tuned and optimized to support the demands of the massively popular [Batman: Arkham Underworld and other franchises]. Turbine invent and simplify microservices persistence services – consolidating their previous NoSQL database solution with highly performant PostgreSQL on Amazon EC2 and Amazon EBS. Turbine also describes other innovative strategies, including integrated analytic techniques to anticipate and predict their scaling operations.
Shreekant Kishor Mandke - Software Development Manager, Amazon.com
In this session, we share how an Amazon.com team that owns a document management platform that manages billions of critical customer documents for Amazon.com migrated from a relational to a non-relational database. Initially, the service was built as an Oracle database. As it grew, the team discovered the limits of the relational model and decided to migrate to a non-relational database. They chose Amazon DynamoDB for its built-in resilience, scalability, and predictability. We provide a template that you can use to migrate from a relational data store to DynamoDB. We also provide details about the entire process: design patterns for moving from a SQL schema to a NoSQL schema; mechanisms used to transition from an ACID (Atomicity, Consistency, Isolation, Durability) model to an eventually consistent model; migration alternatives considered; pitfalls in common migration strategies; and how to ensure service availability and consistency during migration.
Julien Lepine - Solutions Architect Specialist, Amazon Web Services
From servers to workstations, AWS provides the best place to run your Windows workloads. In this session, we'll discuss the ease of deploying Windows workloads on AWS, and architecting for performance, scalability, security, and cost savings. We will explore the use of AWS Directory Service, the Amazon EC2 Run command, and Windows PowerShell to bootstrap your instances for seamless Microsoft Active Directory integration, application installation, and management. We will walk through an architecture that includes Amazon RDS, Amazon EC2, and Amazon WorkSpaces, and discuss the secure relationships among these services. You will learn how you can use native AWS services as well as the tools you are already familiar with to manage your Windows environment.
Alex Smith - Head of Media & Entertainment Architecture, APAC, Amazon Web Services, Amazon Web Services & Callum Hughes - Amazon Studios Senior Solutions Architect, Amazon Studios
The advent of 4K video has resulted in a huge uptick in resource requirements, which is difficult to scale in a traditional environment. The cloud is a perfect environment for handling problems of this scale; however, there are many unanswered questions around best practices and suitable architectures for dealing with massive, high-quality assets. In this session, we will define problem cases and discuss practical architectural patterns for dealing with these challenges by using AWS services such as Amazon EC2 (graphical instances), Amazon EMR, Amazon S3, Amazon S3 Transfer Acceleration, Amazon Glacier, AWS Snowball, and the new magnetic EBS volumes. The best practices that we'll discuss will also be helpful to architects and engineers who are dealing with non-video data. Amazon Studios will present how, powered by AWS, they solved many of these problems and are able to create, manage, and distribute Emmy Award-winning content.
Tod Golding - Partner Solutions Architect, Amazon Web Services
OK, so you have your SaaS application up and running on AWS and now the real fun begins. The adoption of SaaS and a shared multi-tenant infrastructure brings with it a unique set of challenges that require developers to find new and creative ways to optimize their SaaS applications. For many engineers, this means tackling a whole new realm of questions. How do you optimize tenant experience by tiers? How do you perform real-time performance tuning or centralized tenant policy management? How do you collect and aggregate tenant analytics? How do you implement scalable policy-driven configuration of tenant experience? These and many other challenging topics will be covered in this technical deep dive. Tod Golding, an AWS Solutions Architect specializing in helping AWS partners build SaaS systems, will explore a collection of strategies that can be leveraged to increase the performance, flexibility, and agility of SaaS environments. You will learn how to identify areas where services, tools, and design strategies can be combined to enrich the agility of your SaaS architecture.
Bill Jacobi - Senior Solutions Architect, Amazon Web Services
Deploy, scale, and manage the Microsoft servers on AWS supporting 100K users. This session covers the AWS and Microsoft architectures to stand up Exchange, SharePoint, Lync, SQL Server, and Active Directory for enterprises. We focus on the architecture, automation, and management of Windows resources on AWS, including cloud-first technologies like AWS CloudFormation, Amazon EC2 Simple Systems Manager, and Amazon CloudWatch Logs.
Salman Paracha - Principal Product Manager, Amazon Web Services
Organizations today are striving to provide a more flexible environment for their end users, allowing them access to corporate resources from any device, anytime. At the same time, securing corporate information remains a top priority for CIOs. Traditional solutions such as Virtual Desktop Infrastructure (VDI) offer worker flexibility and security benefits, but they are expensive to purchase, complex to deploy, and they don’t scale well. In this session, you’ll learn how Amazon WorkSpaces combines the benefits of VDI with the economics of the cloud to deliver better value for organizations. We’ll show you just how easy it is to get started, and talk through how customers are using Amazon WorkSpaces today.
BAP: Business Applications
Salman Paracha - Principal Product Manager, Amazon Web Services & Richard Perez - Cloud Infrastructure & EUC Manager EMEA, News UK
Amazon WorkSpaces is a desktop computing service that runs in the cloud, and now supports both monthly and hourly billing. In this session, we help you determine the right billing method for your use cases, show you how to provision Amazon WorkSpaces for monthly or hourly billing, and work through a real-world example. News UK, a media organization that owns a stable of news and media brands, describes how they use Amazon WorkSpaces to solve a unique business need.
Nathan Thomas - General Manager, Amazon WorkSpaces, Amazon Web Services & Greg Smith - AVP, Enterprise Operations, University of Maryland University College & Piotr Zielinski - Director of Information Technology, The George Washington University
Join leaders in higher education in this panel discussion as they share how they use Amazon WorkSpaces to move to desktop-as-a-service (DaaS) to provide faculty, staff, and students, access to applications and information they need in classrooms, research labs, and across campus. Amazon WorkSpaces enables major educational institutions be more agile, improve their security posture, and offer end users a more flexible experience, while meeting stringent compliance requirements and remaining cost effective. Hear directly from the world of higher education about their projects to deliver the next generation of end-user computing in their organizations.
Supreeth Sheshadri - Sr.Manager, EC2 New Services, Amazon Web Services & Deepak Suryanarayanan - Principal Product Manager, Amazon Web Services & Sandra Coleman - Senior Director Cloud Strategy and Architecture, Siemens PLM Software & Danny Sicking - Software Engineering Manager, Siemens
Amazon AppStream 2.0 is an application streaming service that provides users instant access to their applications from anywhere, on any connected device. Enterprises can use Amazon AppStream 2.0 to securely deliver the latest versions of their desktop apps to their users. ISVs can reach more users across more devices, without rewriting applications, by streaming their desktop apps from the cloud to their users’ browsers. In this session, we show you how easy it is to upload your apps to Amazon AppStream and start delivering them to thousands of users. We’ll also learn from Siemens PLM how they are moving their desktop apps to the cloud with Amazon AppStream 2.0, and enabling their enterprise customers greater flexibility and productivity through streamed applications.
Pranesh Ramalingam - Senior Product Manager, Amazon Web Services & Thomas Pare - Software Development Manager, Amazon Web Services
Amazon WorkMail is a secure, managed business email and calendar service with support for existing desktop and mobile email clients. Amazon WorkMail enables users to access email, calendars, and contacts using the Amazon WorkMail web app, Microsoft Outlook, or native iOS or Android email applications. In this session, we’ll show you how to quickly create an Amazon WorkMail organization, take a look at the encryption model employed by Amazon WorkMail for data at rest, set up interoperability between existing on-premises Microsoft Exchange servers and Amazon WorkMail and dive into a host of new features that are now available on Amazon WorkMail.
Pravjit Tiwana - General Manager- Amazon WorkDocs, Amazon Web Services & Jim Groff - CEO, Dokkio, Inc.
File sync and share applications have become a cornerstone of collaboration in modern organizations; securing and managing content is an ongoing challenge. Amazon WorKDocs is a fully managed, secure file storage and sharing service with strong administrative controls and feedback capabilities. In this session, we’ll demonstrate how you can use Amazon WorkDocs to manage files and secure content, and show you how we’re shaping Amazon WorkDocs to be a full-fledged productivity tool for users. We'll also present a technical overview of the WorkDocs SDKs, and how they allow you to use existing IT tools for antivirus protection, data loss prevention, analystics, and more.
Salman Paracha - Principal Product Manager, Amazon Web Services & Jason Cradit - Sr.. Director of Technology, TRC & Joel St. Pierre - Event Systems & Support Manager, Autodesk
Amazon WorkSpaces is a desktop computing service that runs in the cloud, and now offers GPU configurations to support design and engineering applications and three-dimensional modeling. We show you how running these applications on Amazon WorkSpaces graphics bundles, in close proximity to data you already store on AWS, can help you process and visualize the results you need. We discuss the economics of running Amazon WorkSpaces graphics bundles, and demonstrate the experience of running a graphics-intensive application on a GPU-enabled Amazon WorkSpace. We also invite Autodesk (or TRC or ESRi) to discuss how they are using Amazon WorkSpaces graphics bundles in their business.
Erik Thomson - Sr Director, Strategic Verticals and Partnerships, Zayo Group & Angelos Kottas - Senior Director, Product Marketing, Salesforce
Your developers are the most important part of transforming your customer interactions into engaging experiences. Salesforce App Cloud, which brings together Heroku, Force.com and Lightning, abstracts away infrastructure and devops complexity, so you can focus on what matters most: building differentiated experiences through apps. Reducing time to market and letting you iterate fast helps you rise above the competition and build lasting customer relationships. In this session, you hear from Zayo, a leading global communications infrastructure services provider, and how they are leveraging the power of integrating the Salesforce and AWS platforms to deliver highly engaging customer experiences, enhancing developer productivity and driving faster innovation cycles. We spotlight Heroku Connect, which makes it easy to extend and synchronize your customer data between Salesforce and AWS and enhance it in ways that empower your developers to do what they do best: innovate. Session sponsored by Salesforce.
Steve Mueller - Specialist Solutions Architect, Amazon Web Services & Rajen Sheth - Senior Director of Product Management, Google & Ziad Lammam - Director of Product Management, Teradici & Forrest Smith - Product Manager, Neverware
You’ve bet big with Amazon WorkSpaces to remove challenges managing your physical fleet of Macs and PCs. Now what? In this session, we’ll demonstrate how you can deploy a rich cloud-based Windows experience on lightweight hardware to reign in management issues, improve TCO, and be at parity with your traditional environment. We’ll take you through the client device ecosystem – from Zero to thin to Google Chrome and Chromium OS clients – and strengthen your ability to determine the right client device strategy moving forward. Live product demonstrations will be provided as we journal how customers are moving to lightweight devices, and what best practices we’ve learned along the way.
Jeff Ferris - WorkSpaces Specialist Solutions Architect, Amazon Web Services & Justin Bradley - WorkSpaces Specialist Solutions Architect, Amazon Web Services
Amazon WorkSpaces is an enterprise desktop computing service that runs in the cloud. In this deep-dive session, we discuss advanced topics and best practices for deploying Amazon WorkSpaces in your organization. We discuss Amazon VPC design and public endpoints, AWS Directory Service, integrating with your on-premises Microsoft Active Directory, using multi-factor authentication, and monitoring and logging with Amazon CloudWatch metrics. We walk through how to do all this using a combination of the AWS Command Line Interface, the AWS Management Console, and AWS CloudFormation templates.
BDA: Big Data
Ryan Oattes - Enterprise Architect, GE Power & Water & Adam Gantt - Solutions Architect, Matillion
Billions of Rows Transformed in Record Time Using Matillion ETL for Amazon Redshift
Russell Smith - Co-Founder, CTO & CIO, Rainforest QA
With Amazon Mechanical Turk (MTurk), you can leverage the power of the crowd for a host of tasks ranging from image moderation and video transcription to data collection and user testing. You simply build a process that submit tasks to the Mechanical Turk marketplace and get results quickly, accurately, and at scale. In this session, Russ, from Rainforest QA, shares best practices and lessons learned from his experience using MTurk. The session covers the key concepts of MTurk, getting started as a Requester, and using MTurk via the API. You learn how to set and manage Worker incentives, achieve great Worker quality, and how to integrate and scale your crowdsourced application. By the end of this session, you will have a comprehensive understanding of MTurk and know how to get started harnessing the power of the crowd.
Aubrey Wing & Bo Mangels - Sr. Manager, Product Marketing, Salesforce
The Internet of Things (IoT) produces vast quantities of data that promise a deep, always connected view into customer experiences through their devices. In this connected age, the question is no longer how do you gather customer data, but what do you do with all that data. How do you ingest at massive scale and develop meaningful experiences for your customers? In this session, you learn how Salesforce IoT Cloud works in concert with the AWS IoT engine to ingest and transform all of the data generated by every one of your customers, partners, devices, and sensors into meaningful action. You also see how customers are using Salesforce and AWS together to process massive quantities of data, build business rules with simple, intuitive tools, and engage proactively with customers in real time. Session sponsored by Salesforce.
Matt Yanchyshyn - Sr. Manager, Solutions Architecture, Amazon Web Services
Building big data applications often requires integrating a broad set of technologies to store, process, and analyze the increasing variety, velocity, and volume of data being collected by many organizations. In this session, we show how you can build entire big data applications using a core set of managed services including Amazon S3, Amazon Kinesis, Amazon EMR, Amazon Elasticsearch Service, Amazon Redshift, and Amazon QuickSight.
Rahul Bhartia - Principal Solutions Architect, Amazon Web Services & Amit Jain - Sr. Manager BI Platform and Reporting, Fanatics
Data is growing at a quantum scale and one of challenges you face is to enable your users to analyze all this data, extract timely insights from it, and visualize it. In this session, you learn about business intelligence solutions available on AWS. We discuss best practices for deploying a scalable and self-serve BI platform capable of churning through large datasets. Fanatics, the nation’s largest online seller of licensed sports apparel, talks about their experience building a globally distributed BI platform on AWS, that delivers massive volumes of reports, dashboards, and charts on a daily basis to an ever growing user base. Fanatics shares the architecture of their data platform, built using Amazon Redshift, Amazon S3, and open source frameworks like Presto and Spark. They talk in detail about their BI platform including Tableau, Microstrategy, and other tools on AWS to make it easy for their analysts to perform ad-hoc analysis and get real-time updates, alerts, and visualizations. You also learn about the experimentation-based approach that Fanatics adopted to fully engage their business intelligence community and make optimal use of their BI platform resources on AWS.
Anurag Gupta - VP, Aurora, EMR and Redshift,, Amazon Web Services
AWS Glue is a fully managed ETL service that makes it easy to understand your data sources, prepare the data for analytics, and load it reliably to your data stores. In this session, we will introduce AWS Glue, provide an overview of its components, and discuss how you can use the service to simplify and automate your ETL process. We will also talk about when you can try out the service and how to sign up for a preview.
Rahul Pathak - GM, Amazon EMR, Amazon Web Services
Amazon Athena is a new interactive query service that makes it easy to analyze data in Amazon S3, using standard SQL. Athena is serverless, so there is no infrastructure to setup or manage, and you can start analyzing your data immediately. You don’t even need to load your data into Athena, it works directly with data stored in S3.
Vidhya Srinivasan - Director, Software Development, Redshift, Amazon Web Services
In this session, you learn about the latest and hottest features of Amazon Redshift. Join Vidhya Srinivasan, General Manager of Amazon Redshift, to take a deep dive into the architecture and inner workings of Amazon Redshift. You discover how the recent availability, performance, and manageability improvements we’ve made can significantly enhance your end user experience. You also get a glimpse of what we are working on and our plans for the future.
BDM: Big Data Mini-Conference
Siva Raghupathy - Sr. Manager, Solutions Architecture, Amazon Web Services
The world is producing an ever increasing volume, velocity, and variety of big data. Consumers and businesses are demanding up-to-the-second (or even millisecond) analytics on their fast-moving data, in addition to classic batch processing. AWS delivers many technologies for solving big data problems. But what services should you use, why, when, and how? In this session, we simplify big data processing as a data bus comprising various stages: ingest, store, process, and visualize. Next, we discuss how to choose the right technology in each stage based on criteria such as data structure, query latency, cost, request rate, item size, data volume, durability, and so on. Finally, we provide reference architecture, design patterns, and best practices for assembling these technologies to solve your big data problems at the right cost.
Radhika Ravirala - Solutions Architect, Amazon Web Services & Tony Gibbs - Solution Architect, Amazon Web Services
Want to get ramped up on how to use Amazon's big data web services and launch your first big data application on AWS? Join us in this workshop as we build a big data application in real time using Amazon EMR, Amazon Redshift, Amazon Kinesis, Amazon DynamoDB, and Amazon S3. We review architecture design patterns for big data solutions on AWS, and give you access to a take-home lab so that you can rebuild and customize the application yourself.
Scott Donaldson - Senior Director, FINRA & Vincent Saulys - Senior Director, FINRA
Data science is a key discipline in a data-driven organization. Through analytics, data scientists can uncover previously unknown relationships in data to help an organization make better decisions. However, data science is often performed from local machines with limited resources and multiple datasets on a variety of databases. Moving to the cloud can help organizations provide scalable compute and storage resources to data scientists, while freeing them from the burden of setting up and managing infrastructure.
Guy Farber - Business Development Manager, Amazon Web Services & Sergi Mesquida Delgado - Head of Innovation & New Ventures, Hotelbeds Group & Alvaro de Nicolas - CTO, Hotelbeds Group
Amazon QuickSight is a fast BI service that makes it easy for you to build visualizations, perform ad-hoc analysis, and quickly get business insights from your data. QuickSight is built to harness the power and scalability of the cloud, so you can easily run analysis on large datasets, and support hundreds of thousands of users. In this session, we’ll demonstrate how you can easily get started with Amazon QuickSight, uploading files, connecting to S3 and Redshift and creating analyses from visualizations that are optimized based on the underlying data. Once we’ve built our analysis and dashboard, we’ll show you easy it is to share it with colleagues and stakeholders in just a few seconds. And with SPICE – QuckSight’s in-memory calculation engine – you can go from data to insights, faster than ever.
Roger Barga - General Manager, Amazon Web Services & Richard Freeman - Lead Data Engineer and Architect, JustGiving
Join us for this general session where AWS big data experts present an in-depth look at the current state of big data. Learn about the latest big data trends and industry use cases. Hear how other organizations are using the AWS big data platform to innovate and remain competitive. Take a look at some of the most recent AWS big data announcements, as we kick off the Big Data re:Source Mini Con.
Marc Teichtahl - Solutions Architect, Amazon Web Services & Daniel Zoltak - Solutions Architect, Amazon Web Services & Tim Bart - CTO, Hello
The growing popularity and breadth of use cases for IoT are challenging the traditional thinking of how data is acquired, processed, and analyzed to quickly gain insights and act promptly. Today, the potential of this data remains largely untapped. In this session, we explore architecture patterns for building comprehensive IoT analytics solutions using AWS big data services. We walk through two production-ready implementations. First, we present an end-to-end solution using AWS IoT, Amazon Kinesis, and AWS Lambda. Next, Hello discusses their consumer IoT solution built on top of Amazon Kinesis, Amazon DynamoDB, and Amazon Redshift.
Jonathan Fritz - Sr. Product Manager, Amazon Web Services & Yekesa Kosuru - V.P. Engineering, DataXu & Dong Jiang - Software Engineer, DataXu Inc
Organizations need to perform increasingly complex analysis on data — streaming analytics, ad-hoc querying, and predictive analytics — in order to get better customer insights and actionable business intelligence. Apache Spark has recently emerged as the framework of choice to address many of these challenges. In this session, we show you how to use Apache Spark on AWS to implement and scale common big data use cases such as real-time data processing, interactive data science, predictive analytics, and more. We talk about common architectures, best practices to quickly create Spark clusters using Amazon EMR, and ways to integrate Spark with other big data services in AWS. This session will feature DataXu, a provider of programmatic marketing and analytics software. DataXu will share how they architected their petabyte-scale ETL processing pipeline and data science workflows using Spark.
Jon Handler - Principal Solutiuons Architect, Amazon Web Services
Elasticsearch is a fully featured search engine used for real-time analytics, and Amazon Elasticsearch Service makes it easy to deploy Elasticsearch clusters on AWS. With Amazon ES, you can ingest and process billions of events per day, and explore the data using Kibana to discover patterns. In this session, we use Apache web logs as example and show you how to build an end-to-end analytics solution. First, we cover how to configure an Amazon ES cluster and ingest data into it using Amazon Kinesis Firehose. We look at best practices for choosing instance types, storage options, shard counts, and index rotations based on the throughput of incoming data. Then we demonstrate how to set up a Kibana dashboard and build custom dashboard widgets. Finally, we dive deep into the Elasticsearch query DSL and review approaches for generating custom, ad-hoc reports.
Richard Freeman - Lead Data Engineer and Architect, JustGiving
Organizations need to gain insight and knowledge from a growing number of Internet of Things (IoT), application programming interfaces (API), clickstreams, unstructured and log data sources. However, organizations are also often limited by legacy data warehouses and ETL processes that were designed for transactional data. Building scalable big data pipelines with automated extract-transform-load (ETL) and machine learning processes can address these limitations. JustGiving is the world’s largest social platform for online giving. In this session, we describe how we created several scalable and loosely coupled event-driven ETL and ML pipelines as part of our in-house data science platform called RAVEN. You learn how to leverage AWS Lambda, Amazon S3, Amazon EMR, Amazon Kinesis, and other services to build serverless, event-driven, data and stream processing pipelines in your organization. We review common design patterns, lessons learned, and best practices, with a focus on serverless big data architectures with AWS Lambda.
Allan MacInnis - Solutions Architect, Amazon Web Services
As more and more organizations strive to gain real-time insights into their business, streaming data has become ubiquitous. Typical streaming data analytics solutions require specific skills and complex infrastructure. However, with Amazon Kinesis Analytics, you can analyze streaming data in real-time with standard SQL—there is no need to learn new programming languages or processing frameworks.
Kurt Brown - Director, Data Platform, Netflix & Eva Tse - Director of Big Data Platform, Netflix
Amazon S3 is the central data hub for Netflix's big data ecosystem. We currently have over 1.5 billion objects and 60+ PB of data stored in S3. As we ingest, transform, transport, and visualize data, we find this data naturally weaving in and out of S3. Amazon S3 provides us the flexibility to use an interoperable set of big data processing tools like Spark, Presto, Hive, and Pig. It serves as the hub for transporting data to additional data stores / engines like Teradata, Redshift, and Druid, as well as exporting data to reporting tools like Microstrategy and Tableau. Over time, we have built an ecosystem of services and tools to manage our data on S3. We have a federated metadata catalog service that keeps track of all our data. We have a set of data lifecycle management tools that expire data based on business rules and compliance. We also have a portal that allows users to see the cost and size of their data footprint. In this talk, we’ll dive into these major uses of S3, as well as many smaller cases, where S3 smoothly addresses an important data infrastructure need. We will also provide solutions and methodologies on how you can build your own S3 big data hub.
Jonathan Fritz - Sr. Product Manager, Amazon Web Services & Naveen Avalareddy - Sr. Principal Architect, Asurion
Amazon EMR is one of the largest Hadoop operators in the world. In this session, we introduce you to Amazon EMR design patterns such as using Amazon S3 instead of HDFS, taking advantage of both long and short-lived clusters, and other Amazon EMR architectural best practices. We talk about how to scale your cluster up or down dynamically and introduce you to ways you can fine-tune your cluster. We also share best practices to keep your Amazon EMR cluster cost-efficient. Finally, we dive into some of our recent launches to keep you current on our latest features. This session will feature Asurion, a provider of device protection and support services for over 280 million smartphones and other consumer electronics devices. Asurion will share how they architected their petabyte-scale data platform using Apache Hive, Apache Spark, and Presto on Amazon EMR.
Eric Ferreira - Principal Database Engineer, Amazon Web Services & Philipp Mohr - Sr. CRM Director, King.com & Pulitha Liyanagama - Technical Director, King.com
Analyzing big data quickly and efficiently requires a data warehouse optimized to handle and scale for large datasets. Amazon Redshift is a fast, petabyte-scale data warehouse that makes it simple and cost-effective to analyze all of your data for a fraction of the cost of traditional data warehouses. In this session, we take an in-depth look at data warehousing with Amazon Redshift for big data analytics. We cover best practices to take advantage of Amazon Redshift's columnar technology and parallel processing capabilities to deliver high throughput and query performance. We also discuss how to design optimal schemas, load data efficiently, and use work load management.
Roy Ben-Alta - Sr. Business Development Manager, Amazon Web Services & Ryan Nienhuis - Senior Product Manager - Technical, Amazon Web Services & Ram Kumar Rengaswamy - Co-Founder and CTO, Beeswax
Amazon Kinesis is a platform of services for building real-time, streaming data applications in the cloud. Customers can use Amazon Kinesis to collect, stream, and process real-time data such as website clickstreams, financial transactions, social media feeds, application logs, location-tracking events, and more. In this session, we first cover best practices for building an end-to-end streaming data applications using Amazon Kinesis. Next, Beeswax, which provides real-time Bidder as a Service for programmatic digital advertising, will talk about how they built a feature-rich, real-time streaming data solution on AWS using Amazon Kinesis, Amazon Redshift, Amazon S3, Amazon EMR, and Apache Spark. Beeswax will discuss key components of their solution including scalable data capture, messaging hub for archival, data warehousing, near real-time analytics, and real-time alerting.
Andre Dufour - Principal Product Manager, Amazon Web Services & Hook Hua - Data Scientist, Jet Propulsion Laboratory
Scaling allows cloud resources to scale automatically in reaction to the dynamic needs of customers. This session will show how Auto Scaling offers an advantage to everyone – whether it’s basic fleet management to keep instances healthy as an EC2 best practice, or dynamic scaling to manage “extremes”. We’ll share examples of how Auto Scaling is helping customers of all sizes and industries unlock use cases and value. We’ll also discuss how Auto Scaling is evolving to scaling different types of elastic AWS resources beyond EC2 instances. NASA Jet Propulsion Laboratory (JPL) / California Institute of Technology will share how Auto Scaling is used to scale science data processing of Interferometric Synthetic Aperture Radar (InSAR) data from earth-observing satellite missions, and reduce response times during hazard response events such as those from earthquakes, floods, and volcanoes. JPL will also discuss how they are integrating their science data systems with the AWS ecosystem to expand into NASA’s next two large-scale missions with remote-sensing radar-based observations. Learn how Auto Scaling is being used at a global scale – and beyond!
Joshua Burgin - General Manager, EC2 Spot, Amazon Web Services
Amazon EC2 provides you with the flexibility to cost optimize your computing portfolio through purchasing models that fit your business needs. With the flexibility of mix-and-match purchasing models, you can grow your compute capacity and throughput and enable new types of cloud computing applications with the lowest TCO.
Sebastian Dreisch - Business Development, Amazon Web Services
Amazon EC2 changes the economics of computing and provides you with complete control of your computing resources. It is designed to make web-scale cloud computing easier for developers. In this session, we will take you on a journey, starting with the basics of key management and security groups and ending with an explanation of Auto Scaling and how you can use it to match capacity and costs to demand using dynamic policies. We will also discuss tools and best practices that will help you build failure resilient applications that take advantage of the scale and robustness of AWS regions
Chunky Gupta - Software Engineer, Yelp & David Morrison - Software Engineer, Yelp
Over the last year, Yelp has transitioned its scalable and reliable parallel task execution system, Seagull, from On-Demand and Reserved Instances entirely to Spot Fleet. Seagull runs over 28 million tests per day, launches more than 2.5 million Docker containers per day, and uses over 10,000 vCPUs in Spot Fleet at peak capacity. To deal with rising infrastructure costs for Seagull, we have extended our in-house Auto Scaling Engine called FleetMiser to scale the Spot Fleet in response to demand. FleetMiser has reduced Seagull’s cluster costs by 60% in the past year and saved Yelp thousands of dollars every month.
Georgios Elissaios - Principal Product Manager, Amazon Web Services
Amazon Lightsail is the latest addition to the AWS family of compute services and the fastest way to get your next cloud server up and running. For a low price that starts at $5/month, Lightsail offers a bundle of resources and services that let you jumpstart your cloud project in a few clicks. The new, intuitive Lightsail console makes it simple to manage your virtual resources, letting you focus on code, not system administration. Come to this session and learn how Lightsail can get you started on AWS quickly and efficiently.
David Pellerin - Business Development Principal, Amazon Web Services & Pieter van Rooyen - CEO, Edico Genome
High performance computing in the cloud is enabling high scale compute- and graphics-intensive workloads across industries, ranging from aerospace, automotive, and manufacturing to life sciences, financial services, and energy. AWS provides application developers and end users with unprecedented computational power for massively parallel applications, in areas such as large-scale fluid and materials simulations, 3D content rendering, financial computing, and deep learning. This session provides an overview of HPC capabilities on AWS, describes the newest generations of accelerated computing instances (including P2), as well as highlighting customer and partner use-cases across industries.
Danielle Greshock - Manager, Solutions Architecture, Amazon Web Services
AWS is an elastic, secure, flexible, and developer-centric ecosystem that serves as an ideal platform for Docker deployments. AWS offers the scalable infrastructure, APIs, and SDKs that integrate tightly into a development lifecycle and accentuate the benefits of the lightweight and portable containers that Docker offers to its users.
Jeremy Edberg - Founder and CEO, Minops
Serverless architectures let you build and deploy applications and services with infrastructure resources that require zero administration. In the past, you had to provision and scale servers to run your application code, install and operate distributed databases, and build and run custom software to handle API requests. Now, AWS provides a stack of scalable, fully-managed services that eliminates these operational complexities.
Paul Andrushkiw & MJ DiBerardino - CTO, Cloudnexa
A worldwide dynamic website infrastructure at the cost and simplicity of a static deployment was once an unachievable dream. That’s no longer the case with the tools and services that AWS offers like Amazon CloudFront, Amazon S3, Amazon API Gateway and AWS Lambda. During this session we will discuss a DevOps approach to managing your code repositories, code quality, continuous integration and deployments. We discuss some static website generators such as mkdocs, Jekyll, and middleman for building knowledge bases, blogs and websites in a serverless architecture.
Peter Norton - Ops Engineer, Librato & Mike Heffner - Data Engineer, SolarWinds
At Librato, a Solarwinds company, we run hundreds of Cassandra instances across multiple rings and use it as our primary data store. In the past year, we embarked on a process to upgrade our fleet of Cassandra Amazon EC2 instances from instance store to instances using Amazon EBS and attached elastic network interfaces (ENIs). We find running Cassandra on EBS gives us the flexibility to choose the best instances for the best performance of our workload while saving us significant costs on infrastructure. In this session, we discuss how Librato operates Cassandra on EBS. Topics include how we chose the right instance for our workload, use detached EBS volumes and ENI mobility to reduce MTTR, use mixed EBS storage types for the best cost/performance tradeoff, debug performance issues, and continuously monitor Cassandra to get the most from AWS. We also look at performance tradeoffs made in the implementation of storage engines of large data systems like Cassandra.
Adam Boeglin - HPC Solutions Architect, Amazon Web Services
Amazon EC2 provides a broad selection of instance types to accommodate a diverse mix of workloads. In this session, we provide an overview of the Amazon EC2 instance platform, key platform features, and the concept of instance generations. We dive into the current generation design choices of the different instance families, including the General Purpose, Compute Optimized, Storage Optimized, Memory Optimized, and GPU instance families. We also detail best practices and share performance tips for getting the most out of your Amazon EC2 instances.
Anu Sharma - Principal Product Manager, Amazon Web Services & Gaurav Agarwal - Software Engineer, AOL Inc & Durga Nemani - System Architect, AOL & Charles Allen - Sr. Software Engineer, Metamarkets
Amazon EC2 Spot instances provide acceleration, scale, and deep cost savings to run time-critical, hyper-scale workloads for rapid data analysis. In this session, AOL and Metamarkets will present lessons learned and best practices from scaling their big data workloads using popular platforms like Presto, Spark and Druid.
Luke Hoban - Product Manager, Amazon Web Services & Scott Burks - Senior Application Engineer, Pearson
In this session, customers learn more about the T2 instance type and the performance and cost savings it can bring to startups, SMBs, and enterprises. Customers will share best practices and tips for how they use T2 instances across workloads including development and test, production web servers, continuous integration and more.
Mark Nunnikhoven - Vice President, Cloud Research, Trend Micro
Each of the four distinct compute services available from AWS bring unique advantages to your application's design: Amazon EC2, AWS Elastic Beanstalk, Amazon ECS, and AWS Lambda. However, your security responsibilities change with each of these services. For example, with Amazon EC2 and AWS Elastic Beanstalk, you need a plan to lock down the operating system, your applications, and your data. Containers running in Amazon ECS provides additional flexibility and mobility but also introduces new challenges as most security techniques were not designed with containers in mind. AWS Lambda requires a completely new approach to security design at the function level.
Rahul Bhartia - Principal Solutions Architect, Amazon Web Services & Ion Stoica - Co-Founder and Professor, Databricks and UC Berkeley
Apache Spark is well known across industries, use cases and businesses of all sizes for its speed and ease of use in sophisticated analysis of large datasets. In this session, learn from Ion Stoica who co-led the Apache Spark project at the AMPLab (UC Berkeley) and co-founder of Databricks, about some of the latest innovations in Spark 2.0, a new open source tool Earnest to choose the optimal cluster configuration for your job, and how and why Databricks choose EC2 to run Spark. We’ll also take a look at how Amazon EC2 and the latest enhancements enable Sparks as a data processing platform, along-with best practices and cost optimization techniques for using Spark with AWS.
Boyd McGeachie - Business Development, Amazon Web Services & Will White - CIO, Mapbox & Nicholas Silva - Infrastructure Lead, IFTTT
Amazon EC2 allows you to bid for and run spare EC2 capacity, known as Spot instances, in a dynamically priced market. On average, customers save 80% to 90% compared to On Demand prices by using Spot instances. Achieving these savings has historically required time and effort to find the best deals while managing compute capacity as supply and demand fluctuate.
Mikhail Prudnikov - Senior Solutions Architect, Amazon Web Services & Robert Tarrall - Director of DevOps, ProtectWise, Inc & Gene Stevens - Co-Founder & CTO, ProtectWise
ProtectWise has been hailed as one of the top 10 coolest startups of 2016, disrupting the security industry by providing cutting edge cloud-based network threat detection, forensics and analytics for Fortune 2000 customers. By developing a scalable, stateless platform, they are able to quickly right-size Amazon EC2 instances to optimize performance and ingest gigabytes of network packet capture at a fraction of the cost. Learn how their cloud-first approach towards infrastructure and application development has enabled them to be more agile, grow faster, and save money. In this session, attendees gain practical best practices for how to build cloud-ready applications and see real-life examples of how this hot startup applied those to common workloads like Cassandra.
Puneet Agarwal - Enterprise Solutions Architect, Amazon Web Services & Vinay Polavarapu - Software Architect/Lead, Verizon & Christian Egeler - Director of AR/VR Product, Verizon
In six months, Verizon has built a best-in-class Augmented Reality and Virtual Reality (AR/VR) platform that streams HD video and game experiences using Amazon EC2 GPU Accelerated instances and CloudFront. Verizon will share their reference architecture and configuration best practices that enabled them to develop a massively scalable VR architecture that scales to support for 100K simultaneous HD video streams to customers around the globe.
Steven Bryen - Manager, Solutions Architecture & Francois de Bodinat - CMO, ZeroLight
AWS GPU computing capabilities have allowed hot start-ups like Zerolight to revolutionize the car buying experience for clients like Audi, by providing advanced 3D car rendering simulations for their customers interested in personalizing cars in real-time at dealer showrooms and online. Zerolight chose to build their company from the ground up on AWS because of GPU-compute capabilities, scale, elasticity, and reliability, allowing customers to perform complex renderings quickly and in real-time. In this session, learn how Zerolight has implemented a scalable, cost-effective, highly responsive 3D rendering platform using Amazon EC2 G2 instances.
Tom 'Elvis' Jones - Partner Solutions Architect, Amazon Web Services & Diego Oppenheimer - CEO, Algorithmia
Algorithmia is a startup with a mission to make state of the art machine learning discoverable by everyone - they offer the largest algorithm marketplace in the world, with over 2500 algorithms supporting tens of thousands of application developers. Algorithma is the first company to make deep learning, one of the most conceptually difficult areas of computing, accessible to any company via microservices. In this session, you learn how this startup has selected and optimized Amazon EC2 instances for various algorithms (including the latest generation of GPU optimized instances), to create a flexible and scalable platform. They also share their architecture and best practices for getting any computationally-intensive application started quickly.
Nicholas Matthews - Partner Solutions Architect, Amazon Web Services
Many customers are using Amazon EC2 instances to run applications with high performance networking requirements. In this session, we provide an overview of Amazon EC2 network performance features (enhanced networking, ENA, placement groups, etc.), and discuss how we are innovating on behalf of our customers to improve networking performance in a scalable and cost-efficient manner. We share best practices and performance tips for getting the best networking performance out of your Amazon EC2 instances.
Bob Griffiths - Solutions Architect Manager, Amazon Web Services & Brett Shriver - Sr. Director Market Regulation Technology, FINRA
FINRA is a leader in the Financial Services industry who sought to move toward real-time data insights of billions of time-ordered market events by migrating from SQL batch processes on-prem, to Apache Spark in the cloud. By using Apache Spark on Amazon EMR, FINRA can now test on realistic data from market downturns, enhancing their ability to provide investor protection and promote market integrity (FINRA enacts rules and provides guidance that securities exchanges & brokers must follow). By using AWS Spot instances, FINRA has saved up to 50% from its on premises solution, increased elasticity/scalability, and accelerated reprocessing requests (from months to days). Learn best practices on how FINRA moves toward real-time data analytics with Spark and AWS, while managing production workloads in parallel, increasing performance and IT efficiency, reducing cost, and modernizing and scaling their infrastructure to prepare for real-time processing in the future.
Letian Feng - Sr. Product Mgr - Technical, Amazon Web Services & POUYAN DJAHANI - Director, Aon Benfield & Oliver Gunasekara - CEO & Co-founder, NGCodec
Accelerated computing is on the rise because of massively parallel, compute-intensive workloads such as deep learning, 3D content rendering, financial computing, and engineering simulations. In this session, we provide an overview of our accelerated computing instances, including how to choose instances based on your application needs, best practices and tips to optimize performance, and specific examples of accelerated computing in real-world applications.
Nathan McGuirt - Manager, Solutions Architecture, Amazon Web Services & Gabriele Garzoglio - Gabriele Garzoglio, HEP Cloud Facility Project Manager, Department of Energy Fermi National Accelerator Laboratory
Every day, the computing power of high-performance computing (HPC) clusters helps scientists make breakthroughs, such as proving the existence of gravitational waves and screening new compounds for new drugs. Yet building HPC clusters is out of reach for most organizations, due to the upfront hardware costs and ongoing operational expenses. Now the speed of innovation is only bound by your imagination, not your budget. Researchers can run one cluster for 10,000 hours or 10,000 clusters for one hour anytime, from anywhere, and both cost the same in the cloud. And with the availability of Public Data Sets in Amazon S3, petabyte scale data is instantly accessible in the cloud. Attend and learn how to build HPC clusters on the fly, leverage Amazon’s Spot market pricing to minimize the cost of HPC jobs, and scale HPC jobs on a small budget, using all the same tools you use today, and a few new ones too.
Andy Katz - Sr. Product Manager, Amazon Web Services & Manuel Pata - CloudOps Engineer, OutSystems SA
AWS Step Functions is a new, fully managed service that makes it easy to coordinate the components of distributed applications and microservices using visual workflows. Step Functions is a reliable way to coordinate components and step through the functions of your application. A graphical console helps you arrange and visualize the components of your application as a series of steps. Step Functions automatically triggers and tracks each step and retries when there are errors so that your application executes in order―and as expected―every time. This session shows how to use Step Functions to create, run, and debug multi-service applications in a matter of minutes. We also share how customers are using Step Functions to reliably build and scale multi-step applications such as order processing, report generation, and data transformation―and to innovate faster.
John Phillips - Sr. Manager, Amazon EC2, Amazon Web Services & Supreeth Sheshadri - Sr.Manager, EC2 New Services, Amazon Web Services & Tosh Tambe - Head of Integration - NICE srl, Amazon Web Services
AWS provides unprecedented computational power for graphics-intensive applications in areas such as design, engineering simulations, and 3D content rendering. Together, Amazon EC2 Elastic GPUs and Amazon AppStream 2.0 provide the capabilities necessary for end users to access and run these applications. In this session, you learn more about Elastic GPUs and Amazon AppStream 2.0, and how you can run graphics-intensive applications on AWS. You also hear from ANSYS, a leader in engineering simulation software, and why they are moving the ANSYS Enterprise Cloud to Elastic GPUs and Amazon AppStream 2.0 to deliver a better experience for customers.
Supreeth Sheshadri - Sr.Manager, EC2 New Services, Amazon Web Services & Mats Westenius - Head of Platform, AVEVA AB
Amazon AppStream 2.0 is an application streaming service that provides users with instant access to their applications from anywhere, on any connected device. Delivering desktop apps from the cloud allows enterprises to manage user access easily and keep apps current; ISVs can reach more users across more devices. In this session, we cover best practices for streaming desktop applications from the AWS Cloud with Amazon AppStream 2.0, and show you how you can address advanced topics such as user authentication, security, and managing an application catalog.
Steven Jones - Sr Manager, Solutions Architecture, Amazon Web Services & Rudi Leibbrandt - Senior Director, HANA Product Management, SAP
AWS and SAP have worked together closely to certify the AWS platform so that companies of all sizes can fully realize all the benefits of the SAP HANA in-memory database platform on the AWS cloud. By placing SAP systems in the cloud, organizations are achieving greater agility, flexibility, and cost efficiency while saving resources to focus on their core businesses. We will discuss recent SAP and AWS innovations including the Amazon EC2 X1 instance type that offers up to 2TB of RAM, and dive into features of the AWS platform that bring significant flexibility to SAP HANA deployments.
Jamie Kinney - Principal Product Manager, Amazon Web Services & Dougal Ballantyne - Principal Product Manager, Amazon Web Services
AWS Batch is a fully-managed service that enables developers, scientists, and engineers to easily and efficiently run batch computing workloads of any scale on AWS. AWS Batch automatically provisions compute resources and optimizes the workload distribution based on the quantity and scale of the workloads. With AWS Batch, there is no need to install or manage batch computing software, allowing you to focus on analyzing results and solving problems. AWS Batch plans, schedules, and executes your batch computing workloads across the full range of AWS compute services and features, such as Amazon EC2, Spot Instances, and AWS Lambda. AWS Batch reduces operational complexities, saving time and reducing costs. In this session, Principal Product Managers Jamie Kinney and Dougal Ballantyne describe the core concepts behind AWS Batch and details of how the service functions. The presentation concludes with relevant use cases and sample code.
CON: Containers Mini-Conference
Paul Maddox - Solutions Architect & Alfonso Acosta - Software Engineer, Weaveworks
Managing the infrastructure behind your ECS cluster is an important and crucial task. In this session we will dive deep into best practices to setup and maintain your ECS cluster with tools such as CloudFormation and Auto Scaling. We will discuss how to setup monitoring and logging infrastructure for your cluster and containers. We will also detail how to tighten the security of your container instances with tools such as SELinux. We will do a live deployment of a complete microservices-based ecommerce website, using Weave Net and Weave Scope for container networking, automated service discovery, monitoring and real-time app visualization. We’ll use CloudFormation, Auto Scaling Groups and other services to show how easy it is to architect and deploy a real application on Amazon ECS with fewer moving parts. We’ll also visualize our application using a real-time map of the containers.
Danielle Greshock - Manager, Solutions Architecture, Amazon Web Services & Tim Secor - Engineering Manager, Okta & Jon Todd - Chief Architect, Okta
Keeping consistent environments across your development, test, and production systems can be a complex task. Docker containers offer a way to develop and test your application in the same environment in which it runs in production. You can use tools such as the ECS CLI and Docker Compose for local testing of applications; Jenkins and AWS CodePipeline for building and workflow orchestration; Amazon EC2 Container Registry to store your container images; and Amazon EC2 Container Service to manage and scale containers. In this session, you will learn how to build containers into your development workflow and orchestrate container deployments using Amazon ECS. You will hear how Okta runs 30,000 tests per developer commit and releases 10,000 new lines of code each week to production with a CI system based on 100% AWS services. We'll also discuss how Okta uses ECS for parallelized testing in CI and for production microservices in a multi-region, always on cloud service.
Brandon Chavis - Solutions Architect, Amazon Web Services
Managing and scaling hundreds of containers is a challenging task. A container management solution takes care of these challenges for you, allowing you to focus on developing your application. In this session, we cover the role and tasks of a container management solution and we analyze how four common container management solutions - Amazon EC2 Container Service, Docker for AWS, Kubernetes, and Apache Mesos - stack against each other. We also see how you can easily get started with each of these solutions on AWS.
Anthony Suarez - Head of Engineering, Amazon ECS, Amazon Web Services
As customers build and run production microservices architectures based on containers, having powerful tools to manage the placement and scheduling of these workloads is critical. Amazon ECS allows customers to focus on building their application and removes the need for managing the cluster management software entirely. This session will focus on the capabilities of the new ECS task placement engine, the built-in ECS schedulers, and explore how customers can build custom schedulers using the ECS event stream and Blox, a newly announced open source scheduling framework that gives developers additional choice and control in how they manage containers in production.
Tom Fuller - Sr. Solutions Architect - Strategic Accounts, Amazon Web Services & Jeremy Brown - Lead Engineer, HERE Maps & Pascal Hahn - Director, Cloud Architecture & Engineering, HERE
Through a combination of Amazon ECS and open source technologies, customers are able to build portable CI/CD pipelines on AWS. As container based deployments become more complex, they require additional rigging for integration. In this session, we show how popular Apache products like Kakfa, Storm, and Zookeeper are being deployed on top of Amazon ECS. We hear from HERE, a provider of mapping data, technologies, and services to the automotive, consumer, and enterprise sectors about an approach that leverages Consul from Hashicorp and Amazon ECS clusters for short-cycle deployments and tag-based environment promotion.
Nathaniel Slater - Senior Manager, Solution Architecture, Amazon Web Services & Nick Elser - Director of Engineering, Instacart
Running and managing large scale applications with microservices architectures is hard and often requires operating complex container management infrastructure. Amazon EC2 Container Service (ECS) is a highly scalable, high performance container management service that supports Docker containers and allows you to easily run applications on a managed cluster of Amazon EC2 instances. In this session, we will walk through a number of patterns used by our customers to run their microservices platforms. We will dive deep into some of the challenges of running microservices, such as load balancing, service discovery, and secrets management, and we’ll see how Amazon ECS can help address them. We'll also hear from Instacart how they use a blue/green deployment process to deploy services to ECS and how they manage configuration with a RDS-based metadata service.
Asha Chakrabarty - Senior Solutions Architect, Amazon Web Services & Will White - CIO, Mapbox
Batch computing is a common way for developers, scientists and engineers to run a series of jobs on a large pool of shared compute resources, such as servers, virtual machines, and containers. Amazon ECS makes it easy to run and manage Docker-enabled applications across a cluster of Amazon EC2 instances. In this session will show you how to run batch jobs using Amazon ECS and together with other AWS services, such as AWS Lambda and Amazon SQS. We will see how you can leverage Amazon EC2 Spot Instances to power your ECS cluster and easily scale your batch workloads. You'll hear from Mapbox on how they use ECS to power their entire batch processing architecture to collect and process over 100 million miles of sensor data per day that they use for powering their maps. Mapbox will also discuss how they optimize their batch processing framework on ECS using Spot Instances and demo their open source framework that will help you get up and running with ECS in minutes.
Sathiya Shunmugasundaram - Sr Mgr Enterprise Architecture, Capital One & Gnani Dathathreya - Director, Enterprise Applied Architecture, Capital One
At Capital One, we are using Docker and container technologies to advance microservices adoption, increase efficiencies of cloud resources, and decouple the application layer from the underlying infrastructure. Capital One is a federated organization with a “you build it, you own it” culture that provides autonomy and speed for delivery teams. Each federated team runs and operates their container management stack. In order for the federated teams to accelerate their cloud and container-based apps adoption, we created self-service automation tools for creation and operations management of container management stack.
Hemanth Jayaraman - Sr. Director, DevOps, Rent-A-Center & Aater Suleman - CEO & Co-Founder, Flux7 Labs Inc.
Rent-A-Center’s challenge was to architect, deploy, and manage a mission-critical SAP Hybris ecommerce platform that could scale to 2 million users a month. Together with Flux7, an AWS Advanced Consulting Partner, Rent-A-Center created an AWS-based approach that would help deliver the solution to market faster, in a secure, highly available, PCI-compliant fashion. In this session, we walk through the implementation details of this solution and its challenges, and explore how Rent-A-Center is now able to achieve ROI through agility, scalability, security, and cost savings.
Andrew Spyker - Senior Software Engineer, Netflix
Customers from over all over the world streamed forty-two billion hours of Netflix content last year. Various Netflix batch jobs and an increasing number of service applications use containers for their processing. In this session, Netflix presents a deep dive on the motivations and the technology powering container deployment on top of Amazon Web Services. The session covers our approach to resource management and scheduling with the open source Fenzo library, along with details of how we integrate Docker and Netflix container scheduling running on AWS. We cover the approach we have taken to deliver AWS platform features to containers such as IAM roles, VPCs, security groups, metadata proxies, and user data. We want to take advantage of native AWS container resource management using Amazon ECS to reduce operational responsibilities. We are delivering these integrations in collaboration with the Amazon ECS engineering team. The session also shares some of the results so far, and lessons learned throughout our implementation and operations.
Chad Schmutzer - Solutions Architect, Amazon Web Services
Deep learning is an implementation of machine learning that uses neural networks to solve difficult and complex problems, such as computer vision, natural language processing, and recommendations. Due to the availability of deep learning libraries and frameworks, developers have the ability to enhance the capabilities of their applications and projects. In this workshop, you learn how to build and deploy a powerful deep learning framework called MXNet on containers. The portability and resource management benefit of containers means developers can focus less on infrastructure and more on building.
Asif Khan - Solutions Architect, Amazon Web Services
In this workshop, you learn how to develop a mobile front-end using Swift, and develop a Swift microservices-based web application to be deployed on Amazon ECS.
Deepak Singh - General Manager, Container Services, Amazon Web Services
Join us to learn about the latest developments from Amazon ECS and the container ecosystem. Deepak Singh, General Manager of AWS Container Services, discusses the evolution of containers on AWS and shares our vision for continued innovation in this space. You also hear about how other companies are using the AWS container platform to innovate and build new businesses.
Scott Windsor - Sr. Software Development Engineer, AWS & Mao Geng - Site Reliability Engineer, Pinterest
Are you struggling with bulky images or slow push and pull times? In this session we will walk through the anatomy of a Docker image and provide techniques you can use to optimize images for faster pushes and pulls and reduce your overall storage footprint. We will discuss Docker image building (build containers versus runtime containers to remove unnecessary software), Docker image composition (minimizing the number of layers), the Docker Remote API (optimizing how images are pushed and pulled), and CI/CD Integration (automate building, versioning, and deploying images to production). We’ll also examine the tools that ECR provides to make Docker image management easier so that you can focus on building your application. Finally, we'll hear from Pinterest about how they use ECR and Docker, some valuable experiences gained along the way, and best practices for using ECR with Apache Mesos.
Michael Capicotto - Solutions Architect, Amazon Web Services & Henrik Johansson - Security Solutions Architect, Amazon Web Services
Containers have had an incredibly large adoption rate since Docker was launched, especially from the developer community, as it provides an easy way to package, ship, and run applications. Securing your container-based application is now becoming a critical issue as applications move from development into production. In this session, you learn ways to implement storing secrets, distributing AWS privileges using IAM roles, protecting your container-based applications with vulnerability scans of container images, and incorporating automated checks into your continuous delivery workflow.
CTD: Content Delivery
Jesse Rosenzweig - CTO Elemental, Elemental - Amazon Web Services & Euan McLeod - Global Head of Live Streaming
Traditional broadcast infrastructures had to be built with on-premises architectures to provide five nines of reliability and to ensure the quality of service expected by premium content owners. However, robust live linear streaming services built on AWS are now feasible and in fact can be 100% available. Learn how broadcast media workflows with Elemental Cloud can provide ingest of heterogeneous video sources, fault tolerance across multiple Availability Zones, time synchronization of video streams, and sustained peak workloads in 24/7 applications.
Khawaja Shams - VP of Engineering, Elemental, Amazon Web Services & Alex Zhang - Sr. Product Manager, Amazon Web Services & Mark Lee
The flexibility of AWS provides many advantages in the processing and delivery of bursty video applications such as video on demand and live events. However, full channel playout and advanced monetization have been the domains of traditional on-premises workflows. Learn how Elemental and AWS are architecting media-centric playout, monetization, and delivery services on top of AWS with live and file playlist integration, static and motion graphic overlays, ad content transcoding, and personalized, server-side ad insertion to bring additional video center operations to the cloud.
Vivek Bhat - Sr. Product Manager, Amazon Web Services & Stephen Godwin - Senior Technical Architect, British Broadcasting Corporation (BBC) & Theodore Kim - Head of DevOps Engineering, GoPro
In this session, you’ll hear different customers’ experiences bringing video workloads to AWS. Customer speakers from both traditional broadcasters and cloud-first technology companies will discuss what was same and what was different as they moved both video on demand asset creation and live feed production to AWS for the first time, and how their platforms have evolved with experience using Elastic Transcoder and Elemental Cloud.
Nathan Dye - Software Development Manager, Amazon Web Services
Whether you are building a secure ecommerce application or developing games, security is a key consideration when architecting your application. In this session, you will learn about edge termination of your end user requests and will dive deep into advanced protocols and ciphers, enforcing end-to-end HTTPS connections with AWS Certificate Manager, access control with AWS WAF.
Tom Witman - Head, Business Development Edge Services, Amazon Web Services
End users expect to be able to view static, dynamic, and streaming content anytime, anywhere, and on any device. Amazon CloudFront is a web service that accelerates delivery of your websites, APIs, video content, or other web assets to end users around the globe with low latency, high data transfer speeds, and no commitments. In this session, learn what a content delivery network (CDN) such as Amazon CloudFront is and how it works, the benefits it provides, common challenges and needs, performance, recently released features like HTTP/2 and IPV6 support, pricing, and examples of how customers are using CloudFront.
Nihar Bihani - Sr. Manager, Product Management, Amazon Web Services
This session introduces Lambda@Edge, a new AWS Lambda feature that allows developers to perform simple computations at AWS edge locations in response to CloudFront events. This will be of interest to developers who want to build low-latency, customized web experiences. We cover product functionality and details of the programming model, and we walk through potential use cases.
Alec Peterson - General Manager, CloudFront, Amazon Web Services
In this series of 15-minute technical flash talks you will learn directly from Amazon CloudFront engineers and their best practices on debugging caching issues, measuring performance using Real User Monitoring (RUM), and stopping malicious viewers using CloudFront and AWS WAF.
Lee Atkinson - Solutions Architect, Amazon Web Services & Alec Peterson - General Manager, CloudFront, Amazon Web Services
In this session, we dive deep into how you can integrate Amazon CloudFront and related services into your application, be agile in developing and adapting the application, and follow best practices when configuring the services to improve security and performance, all while reducing costs. Attend this session and learn how to avoid needless forwarding of headers and cookies, test your application when making changes to the origin, version your configuration changes, monitor usage and automate security, create templates for new distributions, configure SSL/TLS certificates, and more.
Alex Smith - Head of Media & Entertainment Architecture, APAC, Amazon Web Services, Amazon Web Services
At AWS, the availability of our services is non-negotiable. While building our own services, such as Amazon CloudFront, we learn from and develop our own design patterns for high availability. In this session, we review several of these design patterns, and we show how you can implement the patterns in your own services or applications built on top of AWS using services such as Amazon Kinesis, AWS Elastic Beanstalk, or AWS Lambda.
Ian Ward - Engineering, Mapbox
Ian Ward, Platform and Security Engineer from Mapbox, discusses how the AWS global edge network helps improve the availability and performance of delivering hundreds of billions of map tiles to hundreds of millions of end users across the globe on mobile devices, in cars, and over the web. In this session, Ian shares insights on how Mapbox manages day-to-day edge operations using Amazon CloudFront logs, dashboards, and ad hoc queries, and how Mapbox has configured CloudFront with dozens of behaviors and origins to customize their content delivery. Mapbox has grown from using a single AWS region to using several regions, so Ian also explains how his team uses Amazon Route 53 and open source tools to simplify complexity around regional failover, and how Mapbox leverages AWS WAF to deter attacks and abuse.
Glyn Smith - Business Development manager, AWS & Andy Shenkler - EVP and Chief Solutions & Technology Officer, Sony DADC New Media Solutions & Rahul Purini - CTO, Funimation Entertainment
Learn from AWS and Sony DADC how the transformation in content aggregation, management, and delivery are optimized with Sony Ven.ue; a Media as a Service (MaaS) offering. In this session, you learn about the Ven.ue architecture, which uses end-to-end AWS services for content preparation, protection, management, and global content delivery. We dive deep on the integrations with AWS that make Amazon CloudFront a core component of the Ven.ue service. You also hear from a customer of Sony Ven.ue, Funimation, on how they benefit from this platform.
Rao Prahlad - Solutions Architect, Amazon Web Services & Carl Youngblood - Lead Software Engineer, Under Armour
Under Armour implemented cross-region replication with Amazon DynamoDB Streams. Come listen as they share the keys to success.
Pavan Pothukuchi - Principal Product Manager, Amazon Web Services & Laxmikanth Malladi - Principal Architect, Northbay Solutions & Ali Khan - Director of Business Intelligence and Analytics, Scholastic
Amazon Redshift is a fast, simple, cost-effective data warehousing solution, and in this session, we look at the tools and techniques you can use to migrate your existing data warehouse to Amazon Redshift. We will then present a case study on Scholastic’s migration to Amazon Redshift. Scholastic, a large 100-year-old publishing company, was running their business with older, on-premise, data warehousing and analytics solutions, which could not keep up with business needs and were expensive. Scholastic also needed to include new capabilities like streaming data and real time analytics. Scholastic migrated to Amazon Redshift, and achieved agility and faster time to insight while dramatically reducing costs. In this session, Scholastic will discuss how they achieved this, including options considered, technical architecture implemented, results, and lessons learned.
Debanjan Saha - General Manager, Amazon Aurora,, Amazon Web Services & Linda Xu - Principal Architect, data and database, Ticketmaster
Amazon Aurora is a MySQL-compatible relational database engine with the speed, reliability, and availability of high-end commercial databases at one-tenth the cost. This session introduces you to Amazon Aurora, explores the capabilities and features of Aurora, explains common use cases, and helps you get started with Aurora. Debanjan Saha, general manager for Aurora, explains how Aurora differs from other commonly available databases while staying compatible with MySQL and providing a high-end, cost-effective alternative to commercial and open-source database engines. In addition, Linda Xu, data architect at Ticketmaster, walks you through Ticketmaster's journey to Amazon Aurora, starting with evaluation through production migration of a critical Ticketmaster database to Amazon Aurora. Ticketmaster is one of the world's top 10 e-commerce companies and the global market leader in ticketing. In this session, Linda discusses how Aurora lets Ticketmaster provide better services to their fans, customers, and clients, and helps reduce the cost and operational burden while giving greater flexibility to support heavy traffic spikes.
Joseph Fluckiger - Senior Software Architect, ThermoFisher & Eliot Horowitz - Co-Founder and CTO, MongoDB
Mass spectrometry is the gold standard for determining chemical compositions, with spectrometers often measuring the mass of a compound down to a single electron. This level of granularity produces an enormous amount of hierarchical data that doesn't fit well into rows and columns. In this talk, learn how Thermo Fisher is using MongoDB Atlas on AWS to allow their users to get near real-time insights from mass spectrometry experiments—a process that used to take days. We also share how the underlying database service used by Thermo Fisher was built on AWS.
Justin Graham - Sr Manager, Product Management, Amazon Web Services & Aaron Rosenbaum - VP, Strategy, MarkLogic & James Qiu - Director, Engineering, FielMaker & Eddie Lee - VP, Technology, NBC Entertainment
Learn how the AWS Marketplace brings together customers who have challenges with ISVs who have solutions to those challenges. See how to use relational and NoSQL technologies on AWS to build enterprise and consumer apps. NBC used MarkLogic to deliver an award-winning app that can handle high traffic levels and unexpected usage spikes. NBC’s popular, Emmy-winning, “SNL 40” was launched to celebrate the 40th anniversary of Saturday Night Live, and delivers four decades of sketches and performances. Hosted on AWS, the app — as well as a browser-based platform — are powered by the MarkLogic Enterprise NoSQL database. Come learn from the team who collaborated on this project how to run your own database on AWS, and how to integrate with Amazon RDS and other data stores. A world-recognized automotive brand needed to deliver real-time response about their worldwide fleet vehicles. You will learn how they used a combination of AWS services and FileMaker Cloud, (an Apple subsidiary, procured through AWS Marketplace) to deliver high-scale dealer-facing applications.
Kevin Jernigan - Senior Product Manager, Amazon Web Services & Mark Porter - General Manager, Amazon RDS, Amazon Web Services
After we launched Amazon Aurora, a cloud-native relational database with region-wide durability, high availability, fast failover, up to 15 read replicas, and up to five times the performance of MySQL, many of you asked us whether we could deliver the same features - but with PostgreSQL compatibility. We are now delivering a preview of Amazon Aurora with this functionality: we have built a PostgreSQL-compatible edition of Amazon Aurora, sharing the core Amazon Aurora innovations with the object-oriented capabilities, language interfaces, JSON compatibility, ANSI:SQL:2008 compliance, and broad functional richness of PostgreSQL. Amazon Aurora will provide full PostgreSQL compatibility while delivering more than twice the performance of the community PostgreSQL database on many workloads. At this session, we will be discussing the newest addition to Amazon Aurora in detail.
Will Leinweber - Principal Cloud Engineer, Citus Data, Inc
Join the principal engineer of Citus Cloud for a brief overview of Citus, best use cases for it, and a drill down into how it's run and managed as a hosted service on top of AWS. The orchestration of Citus is homegrown, but comes from years of experience of running millions of PostgreSQL databases on top of AWS. Even if you aren't looking to leverage Citus to help you scale out, in this session you'll gain insights applicable to running and managing your stateful services on top of AWS. Citus is a PostgreSQL extension that transforms the database into a distributed, horizontally scalable database. Companies like Cloudflare use Citus to process 40 TB per day. With Citus MX, applications can take advantage of every node in the cluster for writes and yielding near-linear write scaling. Citus MX provide up to 500,000 durable writes per second.
Craig Stires - Head, Big Data and Analytics, APAC, Amazon Web Services
For many organizations, capturing a window of opportunity is their differentiator in the market. They invest in real-time systems to understand and respond to events. With so many use cases, it can be challenging for people to know how to invest in the right platform, and how to build in a cost-optimized way. In this session, we look at how some AWS customers are using real-time analytics to capture windows of opportunity: a telco with a major promotion, an advertising retargeter with global demands, and a personal IOT provider with a lifestyle solution. We dig deeper into their architecture and look for common patterns that can be used to build a real-time analytics platform in a cost-optimized way. We even see how a light-load, real-time analytics system can be built for less than $1000.
Puneet Agarwal - Enterprise Solutions Architect, Amazon Web Services & Steve Abraham - Solutions Architect, Amazon Web Services & Mario Kostelac - Product Engineer, Intercom
Amazon Aurora is a fully managed relational database engine that provides higher performance, availability and durability than previously possible using conventional monolithic database architectures. After launching a year ago, we continued adding many new features and capabilities to Aurora. In this session AWS Aurora experts will discuss the best practices that will help you put these capabilities to the best use. You will also hear from Amazon Aurora customer Intercom on the best practices they adopted for moving live databases with over two billion rows to a new datastore in Amazon Aurora with almost no downtime or lost records.
Greg Khairallah - Business Development Manager, Amazon Web Services & John Winford - Sr. Technical Program Manager, Amazon Web Services & Aaron Carreras - Senior Director, FINRA
You can significantly reduce database licensing and operational costs by migrating from commercial database engines to Amazon Aurora or Amazon RDS for PostgreSQL. In addition to cost reduction you also gain flexibility and operational efficiency by avoiding the frustrating usage constraints that the commercial databases licenses come with. Amazon Aurora and Amazon RDS for PostgreSQL are fully managed database services so you no longer need to worry about complex database management tasks. You can launch a single database instance or thousands of them in just a few minutes, and pay only for what you use. In this session we will dive deep into how AWS Database Migration Service and AWS Schema Conversion Tool help you migrate your commercial databases like Oracle and Microsoft SQL Server to Amazon Aurora or Amazon RDS for PostgreSQL easily and securely with minimal downtime.
Anurag Gupta - VP, Aurora, EMR and Redshift,, Amazon Web Services
Amazon Aurora is a fully managed relational database engine that combines the speed and availability of high-end commercial databases with the simplicity and cost-effectiveness of open source databases. It is purpose-built for the cloud using a new architectural model and distributed systems techniques to provide far higher performance, availability and durability than previously possible using conventional monolithic database architectures. Amazon Aurora packs a lot of innovations in the engine and storage layers. In this session, we will do a deep-dive into some of the key innovations behind Amazon Aurora, new improvements to Aurora's performance, availability and cost-effectiveness and discuss best practices and optimal configurations.
Rick Houlihan - Principal TPM – DBS NoSQL, Amazon Web Services
Explore Amazon DynamoDB capabilities and benefits in detail and learn how to get the most out of your DynamoDB database. We go over best practices for schema design with DynamoDB across multiple use cases, including gaming, AdTech, IoT, and others. We explore designing efficient indexes, scanning, and querying, and go into detail on a number of recently released features, including JSON document support, DynamoDB Streams, and more. We also provide lessons learned from operating DynamoDB at scale, including provisioning DynamoDB for IoT.
Scott Ward - Solutions Architect, Amazon Web Services & KD Singh - Partner Solution Architect, Amazon Web Services & Abdul Sait
Amazon RDS allows customers to launch an optimally configured, secure and highly available database with just a few clicks. It provides cost-efficient and resizable capacity while managing time-consuming database administration tasks, freeing you up to focus on your applications and business. Amazon RDS provides you six database engines to choose from, including Amazon Aurora, Oracle, Microsoft SQL Server, PostgreSQL, MySQL and MariaDB. In this session, we take a closer look at the capabilities of RDS and all the different options available. We do a deep dive into how RDS works and the best practises to achive the optimal perfomance, flexibility, and cost saving for your databases.
Michael Labib - Specialist Solutions Architect, In-Memory NoSQL, Amazon Web Services & Brian Kaiser - CTO, Hudl
In this session, we provide a peek behind the scenes to learn about Amazon ElastiCache's design and architecture. See common design patterns with our Redis and Memcached offerings and how customers have used them for in-memory operations to reduce latency and improve application throughput. During this session, we review ElastiCache best practices, design patterns, and anti-patterns.
Darin Briskman - Technical Evangelist & Developer Outreach, Amazon Web Services & Steve Hunt - Director of Infrastructure, FanDuel & Alan Murray - Director of Architecture, FanDuel & Robin Spira - CTO, FanDuel
Which database is best suited for your use case? Should you choose a relational database or NoSQL or a data warehouse for your workload? Would a managed service like Amazon RDS, Amazon DynamoDB, or Amazon Redshift work better for you, or would it be better to run your own database on Amazon EC2? FanDuel has been running its fantasy sports service on Amazon Web Services (AWS) since 2012. You will learn best practices and insights from FanDuel’s successful migrations from self-managed databases on EC2 to fully-managed database services.
Manish Gupta - CMO, Redis Labs & George Chiramattel Kunjachan - Principal Engineer, Intuit Inc. & Kyle Oppenheim - VP Engineering, Groupon & Jacqueline Hufford-Jensen - Senior Manager, Database Administration, LifeLock, Inc.
Redis Labs' CMO is hosting a fireside chat with leaders from multiple industries including Groupon (e-commerce ), Intuit (Finance ), and LifeLock (Identity Protection ). This conversation-style session will cover the Big Data related challenges faced by these leading companies as they scale their applications, ensure high availability, serve the best user experience at lowest latencies, and optimize between cloud and on-premises operations. The introductory level session will appeal to both developer and DevOps functions. They will hear about diverse use cases such as recommendations engine, hybrid transactions and analytics operations, and time-series data analysis. The audience will learn how the Redis in-memory database platform addresses the above use cases with its multi-model capability and in a cost effective manner to meet the needs of the next generation applications.
Vlad Vlasceanu - Sr. Solutions Architect, Amazon Web Services & Ganesh Subramaniam - Sr. Software Dev Engineer, FBA, Amazon.com & Brandon Cuff - Senior Software Developer, Scopely
We’ll share an overview of leveraging serverless architectures to support high performance data intensive applications. Fulfillment by Amazon (FBA) built the Seller Inventory Authority Platform (IAP) using Amazon DynamoDB Streams, AWS Lambda functions, Amazon Elasticsearch Service, and Amazon Redshift to improve results and reduce costs. Scopely will share how they used a flexible logging system built on Kinesis, Lambda, and Amazon Elasticsearch to provide high-fidelity reporting on hotkeys in Memcached and DynamoDB, and drastically reduce the incidence of hotkeys. Both of these customers are using managed services and serverless architecture to build scalable systems that can meet the projected business growth without a corresponding increase in operational costs.
Radhika Ravirala - Solutions Architect, Amazon Web Services & Nabil Zaman - Software Engineer, Quantcast
Quantcast provides its advertising clients the ability to run targeted ad campaigns reaching millions of online users. The real-time bidding for campaigns runs on thousands of machines across the world. When Quantcast wanted to collect and analyze campaign metrics in real-time, they turned to AWS to rapidly build a scalable, resilient, and extensible framework. Quantcast used Amazon Kinesis streams to stage data, Amazon EC2 instances to shuffle and aggregate the data, and Amazon DynamoDB and Amazon ElastiCache for building scalable time-series databases. With Elastic Load Balancing and Auto Scaling groups, they are able to set up distributed microservices with minimal operation overhead. This session discusses their use case, how they architected the application with AWS technologies integrated with their existing home-grown stack, and the lessons they learned.
Martin Sirull - DevOps Consultant, Amazon Web Services & Jason Chambers - Developer, Toyota Racing Development & Philip Loh - Full Stack Engineer, Toyota Racing Development
Toyota Racing Development (TRD) developed a robust and highly performant real-time data analysis tool for professional racing. In this talk, learn how we structured a reliable, maintainable, decoupled architecture built around Amazon DynamoDB as both a streaming mechanism and a long-term persistent data store. In racing, milliseconds matter and even moments of downtime can cost a race. You'll see how we used DynamoDB together with Amazon Kinesis and Kinesis Firehose to build a real-time streaming data analysis tool for competitive racing.
Padma Malligarjunan - Sr. Technical Account Manager, AWS Enterprise Support, Amazon Web Services & Rohit Dialani - Principal Engineer, DataXu & Yekesa Kosuru - V.P. Engineering, DataXu
'Attribution' is the marketing term of art for allocating full or partial credit to individual advertisements that eventually lead to a purchase, sign up, download, or other desired consumer interaction. We'll share how we use DynamoDB at the core of our attribution system to store terabytes of advertising history data. The system is cost effective and dynamically scales from 0 to 300K requests per second on demand with predictable performance and low operational overhead.
Andrew Baird - Solutions Architect, Amazon Web Services & Chris Taylor - Director, Customer Experience Architecture, Chick-fil-A, Inc.
Chris leads the team providing back-end services for the massively popular Chick-fil-A One mobile app that launched in June 2016. Chick-fil-A follows AWS best practices for web services and leverages numerous AWS services, including Elastic Beanstalk, DynamoDB, Lambda, and Amazon S3. This was the largest technology-dependent promotion in Chick-fil-A history. To ensure their architecture would perform at unknown and massive scale, Chris worked with AWS Support through an AWS Infrastructure Event Management (IEM) engagement and leaned on automated operations to enable load testing before launch.
Liyin Tang - Software Engineer, Airbnb & Greg Brandt - Software Engineer, Airbnb
During this session Greg Brandt and Liyin Tang, Data Infrastructure engineers from Airbnb, will discuss the design and architecture of Airbnb's streaming ETL infrastructure, which exports data from RDS for MySQL and DynamoDB into Airbnb's data warehouse, using a system called SpinalTap. We will also discuss how we leverage Spark Streaming to compute derived data from tracking topics and/or database tables, and HBase to provide immediate data access and generate cleanly time-partitioned Hive tables.
Greg McConnel - Solutions Architect, Amazon Web Services & Zacariah Litton - Vice President of Engineering, Telltale Games
Every choice made in Telltale Games titles influences how your character develops and how the world responds to you. With millions of users making thousands of choices in a single episode, Telltale Games tracks this data and leverages it to build more relevant stories in real time as the season is developed. In this session, you’ll learn about Telltale Games’ migration from Apache CouchDB to Amazon DynamoDB, the challenges of adjusting capacity to handling spikes in database activity, and how it streamlined its analytics storage to provide new perspectives of player interaction to improve its games.
Darin Briskman - Technical Evangelist & Developer Outreach, Amazon Web Services & Nicholas Silva - Infrastructure Lead, IFTTT
IFTTT is a free service that empowers people to do more with the services they love, from automating simple tasks to transforming how someone interacts with and controls their home. IFTTT uses ElastiCache for Redis to store transaction run history and schedule predictions as well as indexes for log documents on S3. Join this session to learn how the scripting power of Lua and the data types of Redis allowed them to accomplish something they would not have been able to elsewhere.
Nathaniel Slater - Senior Manager, Solution Architecture, Amazon Web Services & Benedikt Neuenfeldt - Architect, Sony Interactive Entertainment & Aki Kusumoto - VP, NPS Development Dept., System/Network Engineering & Operation Division, Sony Interactive Entertainment Inc.
In this session, you will learn the key differences between a relational database management service (RDBMS) and non-relational (NoSQL) databases like Amazon DynamoDB. You will learn about suitable and unsuitable use cases for NoSQL databases. You'll learn strategies for migrating from an RDBMS to DynamoDB through a 5-phase, iterative approach. See how Sony migrated an on-premises MySQL database to the cloud with Amazon DynamoDB, and see the results of this migration.
Raju Gulabani - VP, Database Services, Amazon Web Services
Raju Gulabani, vice president of AWS Database Services (AWS), discusses the evolution of database services on AWS and the new database services and features we launched this year, and shares our vision for continued innovation in this space. We are witnessing an unprecedented growth in the amount of data collected, in many different shapes and forms. Storage, management, and analysis of this data requires database services that scale and perform in ways not possible before. AWS offers a collection of such database and other data services like Amazon Aurora, Amazon DynamoDB, Amazon RDS, Amazon Redshift, Amazon ElastiCache, Amazon Kinesis, and Amazon EMR to process, store, manage, and analyze data. In this session, we provide an overview of AWS database services and discuss how our customers are using these services today.
Steve Abraham - Solutions Architect, Amazon Web Services & John Winford - Sr. Technical Program Manager, Amazon Web Services & Abdul Sait & Shree Kenghe - Solutions Architect, Amazon Web Services
It can help you do much more. You can use DMS to consolidate multiple databases into a single database or split a single database into multiple databases. You can also use DMS for data distribution to multiple systems. For both of these use cases your source database can be outside of AWS (on premises) or in AWS (EC2 or RDS). DMS can also be used for near real-time replication of data. Replication can be done to one or more targets within AWS, in the same region or across regions. You can also replicate data from databases within AWS to databases outside of AWS. In this session we will discuss all these usage patterns and help you try them out yourselves.
DAT322: Workshop: Stretching Scalability: Doing more with Amazon Aurora
Amit Sharma - Partner Solutions Architect, Amazon Web Services & KD Singh - Partner Solution Architect, Amazon Web Services & Scott Ward - Solutions Architect, Amazon Web Services
Easy scalability is a powerful feature of Amazon Aurora. Scalability in its actual definition refers to being able to get larger or smaller depending on the need. Amazon Aurora allows you to easily achieve this by scaling the database instance up or down and adding or removing read replicas. Scaling across regions brings additional resilience to your architectures and could boost your application performance due to geographic proximity. You can perform all of these scaling operations through the Aurora console. You can also automate instance and read scaling using lambda function or scripts based on the usage pattern you define. You can extend the automation by feeding your database usage data from Aurora enhanced monitoring into Machine Learning to provide more sophisticated predictive patterns to drive your automation. In this session we will do a deep dive into how scalability works in Aurora and how to make the best use of it to reduce your cost, increase application performance and architect resilient applications.
DAT323: Workshop: Converting Your Oracle or Microsoft SQL Server Database to an Open Source Amazon Aurora or PostgreSQL Database Using AWS SCT and AWS DMS
John Winford - Sr. Technical Program Manager, Amazon Web Services & Raul Frias - Solutions Architect, Amazon Web Services
In this workshop, you migrate a sample sporting event and ticketing database from Oracle or Microsoft SQL Server to Amazon Aurora or Postgre SQL using the AWS Schema Conversion Tool (AWS SCT) and AWS Database Migration Service (AWS DMS). The workshop includes the migration of tables, indexes, procedures, functions, constraints, views, and more. We run SCT on a Amazon EC2 Windows instance--bring a laptop with Remote Desktop (or some other method of connecting to the Windows instance). Ideally, you should be familiar with relational databases, especially Oracle or SQL Server and PostgreSQL or Aurora, to get the most from this session. Additionally, attendees should be familiar with SCT and DMS. Familiarity with SQL Developer and pgAdmin III will be helpful but is not required.
DCS: Developer Community
Austen Collins - CEO, Serverless & Andrew Glover - Engineering Manager, Netflix & Nadia Eghbal - Community Programs, Github
This session discusses the open-source community, its vital place within the AWS ecosystem, and how AWS works to provide seamless integration points. Our speakers share their experiences building and deploying cloud-based open-source projects while also reviewing some of today's most popular and relevant open-source platforms and solutions.
Ryan Kroonenburg - Founder, A Cloud Guru & Paolo Latella - AWS Authorized Instructor, XPeppers & Rich Morrow - Though leader, trainer, mentor in the cloud, big data, IoT & mobile spaces, Global Knowledge
Currently, a breadth of AWS training opportunities are available worldwide, both led by AWS and through community-driven training platforms. In this session, community leaders sort through the different training resources, discuss the resources they used to help them become AWS experts, and explain how different training solutions can complement one another.
Peter Sankauskas - Organizer, Advanced AWS Meetup (SF) & Yu Zhang - Community Manager, BootDev IT Solution Co.,Ltd & Markus Ostertag - Organizer AWS Usergroup Munich, Head of Development Team Internet AG, & Shingo Yoshida - CEO, Section-9, inc.
Our panel of experts lead AWS user groups in San Francisco, Yokohama, Munich, and Shanghai. In this session they share the stories behind how their groups began, list best practices for sustaining a technical meetup over time, and offer advice for AWS enthusiasts who are considering starting up a new user group in their city.
DCS204: Developer Lightning Talks and Happy Hour
Ben Whaley - Cloud consultant, WhaleTech & Greg Wiley - Principal Software Engineer, CJ Affiliate & Anders Bjoernestad - Senior Consultant, Webstep as & Kevin Epstein - CTO, CorpInfo & Minyoung Jeong - CTO, The Beatpacking Company & Teri Radichel - Cloud Architect, WatchGuard Technologies & Patrick Hannah - VP of Engineering, CloudHesive
Don't miss our fast and furious conclusion to the Developer Community Summit. Enjoy food and drinks while listening to 6 community leaders, each of whom shares the latest developer tools, trends, and events taking place in their region in five minutes or less.
Shankar Ramachandran - Solutions Architect, Amazon Web Services & Elaine Tsai - Product Manager, Twilio
In this session we will learn about building Serverless Bots using AWS Services. We will explore contemporary examples of Bots and Serverless architectural building blocks needed to build a Bot. Taking few sample Bots, we will dive deeper into the AWS Services used to build them. The focus will be on Serverless architectural components such as Lambda, API Gateway, Alexa Skills Kit, etc. We will explore different interfaces, voice (eg. Alexa) and text (eg. Slack). We will also discuss about building intelligent Bots.
DEV: DevOps & Tools
Chris Munns - Business Development Manager - DevOps, Amazon Web Services
Today’s cutting edge companies have software release cycles measured in days instead of months. This agility is enabled by the DevOps practice of continuous delivery, which automates building, testing, and deploying all code changes. This automation helps you catch bugs sooner and accelerates developer productivity. In this session, we’ll share the processes followed by Amazon engineers and discuss how you can bring them to your company by using AWS CodePipeline and AWS CodeDeploy, services inspired by Amazon's internal developer tools and DevOps culture.
Ilan Rabinovitch - Director, Technical Community, Datadog & Alan Milford - Systems Architect, AOL
AOL originally provided dial-up service to millions of people. Today, AOL powers advertising and media experiences for the web’s top destinations. How do you maintain observability and reliability to both business and technical teams for high-traffic services in a dynamic infrastructure? Join us as we discuss AOL’s DevOps journey. We will dive into its engineering culture, automation, and monitoring best practices that have allowed AOL to successfully reinvent their infrastructure, as they moved from globally distributed data centers to the AWS Cloud. Session sponsored by Datadog.
Matthew McCullough - Director of Field Services, GitHub, Inc. & Natalie Bradley - Senior Technical Program Manager, Accenture
In this session, you’ll hear from GitHub and Accenture Federal Services, a trusted advisor to the US government, on why they have continued to invest in the adoption of and transition to cloud services. After migrating to AWS cloud, one agency deployed GitHub, the cloud-hosted, distributed version control and collaboration platform, as the backbone of its DevOps program.
Ilan Rabinovitch - Director, Technical Community, Datadog & Matt Williams - Evangelist, Datadog
Just as we got a hang of monitoring our server-based applications, they take away the server. How do you monitor something that doesn’t exist? Which metrics matter most in a serverless world? In this session, we will look at how applications are different in an AWS Lambda-based world and how to monitor them. Join us as we work our way through the stack and demonstrate how to capture the health and performance of your services.
Jeff Hoffman - Director, Elastic Beanstalk, Amazon Web Services
AWS Elastic Beanstalk provides an easy way for you to quickly deploy, manage, and scale applications in the AWS Cloud. Through interactive demos and code samples, this session will teach you how to deploy your code using Elastic Beanstalk, provision and use other AWS services (Amazon SNS, Amazon SQS, and Amazon DynamoDB), use your application’s health metrics to tune performance, scale your application to handle millions of requests, perform zero-downtime deployments with traffic routing, and keep the underlying application platform up-to-date with managed updates. Code samples for demos will be available to all session attendees.
Steven Roberts - Senior Software Development Engineer, Amazon Web Services & Norman Johanson - Senior Software Developer, Amazon Web Services
Just as serverless application development is rapidly becoming the most popular way to bring highly scalable applications to the cloud, .NET has undergone radical changes with .NET Core to become a premier development platform for the cloud. In this session, you will learn how to use the newly launched C# support for .NET Core with AWS Lambda to create highly scalable serverless applications that target platforms from the traditional desktop to mobile devices. We will demonstrate how to write, test, and deploy C# code to AWS Lambda and see how we can leverage our serverless back end from mobile applications.
Dave Hahn - Senior SRE & Chief AWS Botherer, Netflix
Netflix is big. Really big. You just won't believe how vastly, hugely, mind-bogglingly big it is.
Jeremy Oakey - Sr Director, Technical Marketing and Integrations, Cisco Systems
In the digital economy, the fast development and deployment of applications is critical to success. To thrive in this application-oriented business environment, IT organizations are acting now to change their tools and processes to better support agile development methodologies. This session will cover performance benchmarking, benefits of migrating existing workloads, use of key services like Amazon RDS and AWS CloudHSM, and demonstrate how to deploy applications securely and at scale. Session sponsored by Cisco.
Bob Wilkinson - General Manager, Amazon CloudWatch, Amazon Web Services
In this session, we cover three common scenarios that include Amazon CloudWatch Logs and AWS Lambda. First, you learn how to build an Elasticsearch cluster from historical data using Amazon S3, Lambda, and CloudWatch Logs. Next, you learn how to add details to CloudWatch alarm notifications using Amazon SNS and Lambda. Finally, we show you how to bring Elastic Load Balancing logs to CloudWatch Logs using S3 bucket triggers from Lambda.
Armando Leite - Global Security Architect, Amazon Web Services & Prashant Prahlad - Principal Product Manager, Amazon Web Services
AWS CloudTrail, Amazon CloudWatch Events, AWS Identity & Access Management (IAM), Trusted Advisor, AWS Config Rules, other services? In this session, we will help you use existing and recently launched services to automate configuration governance so that security is embedded in the development process. We outline four easy steps (Control, Monitor, Fix, and Audit) and demonstrate how different services can be used to meet your governance needs. We will showcase real-life examples and you can take home a blog post with code examples and the full source code for scripts and tooling that AWS professional services have built using these services.
Theo Carpenter - Systems Manager II, Category Management, Woot & Tom Fuller - Sr. Solutions Architect - Strategic Accounts, Amazon Web Services
In this session, we’ll look at the AWS services that customers are using to build and deploy Microsoft-based solutions that use technologies like Windows, .NET, SQL Server, and PowerShell. We’ll start by showing you how to build a Windows-based CI/CD pipeline on AWS using AWS CodeDeploy, AWS CodePipeline, AWS CloudFormation, and PowerShell using an AWS Quick Start. We’ll also cover best practices for how you can create templates that let you automatically deploy ready-to-use Windows products by leveraging services and tools like AWS CloudFormation, PowerShell, and Git. Woot, an online retailer for electronics, will share how it moved from using a complex mix of custom PowerShell code for its DevOps processes to using services like Amazon EC2 Simple Systems Manager (SSM), AWS CodeDeploy, and AWS Directory Service. This migration eliminated the need for complex PowerShell scripts and reduced the operational complexity of performing operational tasks like renaming servers, joining domains, and securely handling keys.
David Kennedy - Innovation and Development Manager, Dynatrace & Alois Reitbauer - VP, Chief Technology Strategist
At Dynatrace, we challenged ourselves to build a virtual team member to help operations teams run large-scale cloud infrastructures. Think J.A.R.V.I.S. from Iron Man, but for operations. We built our cloud infrastructure on Amazon EC2, Elastic Load Balancing load balancers, and Auto Scaling groups for real-time scalability, Amazon Route 53 for instant customer access, Amazon Echo and Alexa for voice interaction, AWS Lambda for fast prototyping of the human-interaction layer, and Amazon DynamoDB for handling complex conversations. In this session, we will also discuss how we extend the service by using Amazon Machine Learning and AWS IoT to more naturally integrate our virtual assistant into the real world.
Amir Golan - Senior Product Manager, Amazon Web Services
To ensure that your application operates in a predictable manner in both your test and production environments, you must vigilantly maintain the configuration of your resources. By leveraging configuration management solutions, Dev and Ops engineers can define the state of their resources across their entire lifecycle. In this session, we will show you how to use AWS OpsWorks, AWS CodeDeploy, and AWS CodePipeline to build a reliable and consistent development pipeline that assures your production workloads behave in a predictable manner.
Iain Caldwell - Head of Infrastructure, NEWS UK & Joe Kinsella - CTO & Founder, CloudHealth Technologies
When you run a complex AWS environment with thousands of Amazon EC2 instances, more than half a petabyte of object storage, and support the largest daily newspapers in the UK, you need a world-class cloud management strategy. For companies like News Corp, implementing policies that automate infrastructure schedules, right-size workloads, and manage and modify reservations is critical. As you scale your cloud infrastructure, defining centralized governance rules while enabling decentralized management is key to running an optimized cloud.
Mahdi Ben Hamida - Software Engineer, SignalFx
Elasticsearch has quickly become the leading open source technology for scaling search and building document services on. Many software providers have come to rely on it to serve the needs of high-performance, production applications.
James Saryerwinnie - Software Development Engineer, Amazon Web Services
Chalice is a new Python serverless microframework for AWS. In this session, we'll show how to use it to create serverless REST APIs that use Amazon API Gateway and AWS Lambda. We'll deploy our first REST API in less than 30 seconds. Several live demos will show how to build on this example to leverage other Chalice capabilities, including advanced URL routing, automatic policy generation, and Swagger exports.
Julien Lepine - Solutions Architect Specialist, Amazon Web Services
In this session, you will discover how to integrate the AWS developer tools into your development process. We will demonstrate how to leverage AWS services, the .NET SDK, and the Visual Studio Toolkit to simplify and streamline your development processes. This session is targeted at development teams using Microsoft Visual Studio and the Microsoft ecosystem of products. Most of the presentation will be in Visual Studio.
Andrew Troutman - Senior Manager AWS Deployment Services, Amazon Web Services
We’ve seen companies like fast-growing startups, large enterprises, and even Amazon.com itself adopt and evolve strategies to optimize their delivery of cloud applications. Some AWS customers perform in-place updates across their servers. Some perform blue-green deployments to newly provisioned servers. In this session, we’ll share the advantages of each approach and talk about the scenarios in which you should choose one over the other. We will also demonstrate how to perform each deployment technique in an agile and reliable manner.
Andrew Glover - Engineering Manager, Netflix
Netflix rapidly deploys services across multiple AWS accounts and regions over 4,000 times a day. We’ve learned many lessons about reliability and efficiency. What’s more, we’ve built sophisticated tooling to facilitate our growing global footprint. In this session, you’ll learn about how Netflix confidently delivers services on a global scale and how, using best practices combined with freely available open source software, you can do the same.
Lee Atchison - Principal Cloud Architect, New Relic & Brian Lieberman - Systems Engineer, Gannett
Dynamic applications require dynamic resources and dynamic infrastructure. AWS provides many resources for applications to build highly scaled, highly available, dynamic applications, services, and microservices. However, managing and tracking these resources—and making sure they are operating as expected—is a challenge. In this session, we discuss how to monitor and manage the dynamic resources that make up your applications, and learn how to tell when a resource is causing your application problems. Designed for people already acquainted with basic dynamic resource allocation techniques, such as effectively using Auto Scaling, this session helps you take your resource management to the next level.
Anil Kumar - Product Manager, Amazon Web Services & Dominic Divakaruni - Sr.Prod Mgr, Amazon Web Services
In this session, we will review ways to manage the lifecycle of your dev, test, and production infrastructure using CloudFormation. Learn how to architect your infrastructure through loosely coupled stacks using cross-stack references, tightly coupled nested stacks and other best practices. Learn how to use CloudFormation to provision and manage a continuous deployment pipeline for your infrastructure-as-code. Automate deployment of new development environments as your infrastructure evolves, promote your new architecture for testing, and deploy changes to production.
Tung Nguyen - Chief Technical Advisor, Bleacher Report & Benson Wu - DevOps Manager, Bleacher Report
When you're one of the top sports media sites on the internet, you deal with scale like no other. Bleacher Report and its Team Stream app enable millions of users to access their own personalized view of sports. In this session, we'll talk about how we broke up the monolith into microservices and how Elastic Beanstalk empowered us to move quickly. Learn how a small Ops team provided a world-class build/release pipeline by standing on the shoulder of giants (AWS Elastic Beanstalk, Jenkins, and Docker). This session is designed for those who want to get up and running as quickly as possible, but are uncompromising in their ownership of infrastructure. We'll discuss the reasoning behind our switch from fully managing our own infrastructure to a managed service, including some advanced customizations made possible through AWS CloudFormation and AWS Elastic Beanstalk configuration files (.ebextensions).
Abhishek Singh - Principal Product Manager, Amazon Web Services
Analyzing and debugging production distributed applications built using a service-oriented or microservices architecture is a challenging task. In this session, we will introduce AWS X-Ray, a new service that makes it easier to identify performance bottlenecks and errors, pinpoint issues to specific service(s) in your application, identify the impact of issues on users of your application, and visualize a request call graph and service call graph for your applications. We will show interactive demos, and code samples for the demo will be available to all session attendees.
Chayan Biswas - Sr. Product Manager - Technical, Amazon Web Services & Eric Gifford - Information Security Architect, Cambia Health Solutions & Brad Davidson - Security Engineer, Cambia Health Solutions
In this session, we’ll show how customers can use management tools to standardize the creation of AWS resources and then govern these resources through the lifecycle. By using AWS CloudFormation and AWS Service Catalog to provision resources at scale, AWS Config to audit any changes to the configuration of these resources, Amazon CloudWatch to monitor the health of these resources, and AWS CloudTrail to audit who or what made API calls to these resources, customers can automate and scale the administration of their infrastructure on AWS. They can even go one step further and automate compliance checking and remediation by using AWS Config rules and Amazon CloudWatch Events. We will demo how this is possible by looking at some common use cases.
David Lowry - Sr. Cloud Manager, Infor & Amul Merchant - VP, Cloud Architecture and Services, Infor Global
Building scalable automation tools that work across heterogeneous application environments is challenging and can inhibit enterprise cloud migration efforts. Learn how the DevOps team at Infor, one of the world’s leading ISVs, manages dozens of enterprise applications built with a variety of technologies and application architectures. In addition to Infor’s approach to deployment and management automation, this session will cover the core tools they’ve developed on top of native AWS services such as AWS CloudFormation, AWS CodeDeploy, and AWS Lambda.
Adam Jacob - CTO, Chef Software, Inc.
Adam Jacob, the CTO of Chef, will walk through what a DevOps Culture really means, explaining the most important concepts through examples and stories from the trenches. We'll talk about how to build empowered organizations, design delightful products, run effective operations, and more.
Shahbaz Alam - Manager, AWS Professional Services, Amazon Web Services & Peter Marney - SVP and Chief Product Technology Officer, John Wiley and Sons
As incumbent enterprises move to the cloud, questions arise how transform the legacy IT culture to maximize the agility and flexibility AWS provides. Speed and dexterity must be implemented in a consistent manner, minimizing the impact to the organizational structure, but taking into account the existing skill sets and knowledge base. With AWS Service Catalog, you can manage commonly deployed AWS CloudFormation template versions, enable controlled self-provisioning, and leverage those same products in your automated deployment pipelines to AWS. In this session, developers, operations leads, architects, and IT managers learn how to leverage AWS Service Catalog and AWS CloudFormation to transform IT culture to maximize the agility, flexibility, and value that the AWS platform provides. Additionally, John Wiley & Sons, a 200-year-old enterprise, demonstrates how AWS Professional Services helped them balance the velocity achieved by moving to AWS with a structured governance model to deploy their cloud infrastructure and application code.
Sadayuki Furuhashi - Founder & Software Architect, Treasure Data
Learn how to leverage new workflow management tools to simplify complex data pipelines and ETL jobs spanning multiple systems. In this technical deep dive from Treasure Data, company founder and chief architect walks through the codebase of DigDag, our recently open-sourced workflow management project. He shows how workflows can break large, error-prone SQL statements into smaller blocks that are easier to maintain and reuse. He also demonstrates how a system using ‘last good’ checkpoints can save hours of computation when restarting failed jobs and how to use standard version control systems like Github to automate data lifecycle management across Amazon S3, Amazon EMR, Amazon Redshift, and Amazon Aurora. Finally, you see a few examples where SQL-as-pipeline-code gives data scientists both the right level of ownership over production processes and a comfortable abstraction from the underlying execution engines. This session is sponsored by Treasure Data.
Kyle Knapp - SDE, Amazon Web Services
Understanding the internals of the AWS CLI will make you a more effective user. This talk provides a deep dive into the architecture, debugging techniques, advanced usage patterns, and some of its more advanced features. We recommend this talk for people who are already familiar with the AWS CLI because the topics will be advanced. By the end of the talk, audience members will have a deeper understanding of the AWS CLI, allowing them to tackle current or future applications from a different perspective.
Mark Mansour - Sr. Manager, Continuous Delivery, AWS, Amazon Web Services
Continuous delivery makes teams more agile and quickens the pace of innovation. Too often, though, teams adopt continuous delivery without putting the right safety mechanisms in place. In this talk, we'll transform a simple but typical software release process into one that is safe. We'll use DevOps techniques like continuous integration, a variety of non-production testing stages, rollbacks, machine redundancy, Availability Zone redundancy, canary deployments, canary tests, and dashboards. We'll use AWS Lambda, AWS CloudFormation, AWS CodePipeline, AWS CodeDeploy, Amazon CloudWatch alarms and dashboards, and AWS Elastic Beanstalk.
Ganesh Janakiraman , CA & Nicholas Gerasimatos - Red Hat Cloud Evangelist, Red Hat & Jason Dobies - Principal Technical Marketing Engineer, Red Hat
In this session, we walk through the simple process of deploying and managing your own Linux-based application in the cloud and also discuss key use-cases and benefits to automated configuration, deployment, and administration of application stacks.
ENT: Enterprise & Migration
Nick Frank - Practice Lead - Mobility and EUC, AHEAD & Normann Vogel - Senior System Engineer, Informa
Your enterprise has decided it is exiting the traditional desktop business and migrating to Amazon WorkSpaces. Your challenge: how do you provide end users a high quality experience using DaaS while integrating key enterprise services? Focusing on user adoption and simplified operational management DaaS offers significant benefits over traditional physical desktops and VDI solutions. These benefits include hourly consumption pricing, sizing flexibility, linear scalability, and simplified management.
Bill Lynch - VP of Products, Cloudability
In this session, you’ll learn how large, scaled organizations control and optimize their AWS costs. You’ll understand how you can grow usage in a cost-effective way across all business units or product teams. Attendees will walk away with new strategies and tools for driving cost-efficient decision-making at every level of the company.
Pawan Agnihotri - Principal Solutions Architect -- Global Financial Services, Amazon Web Services & Shahbaz Alam - Manager, AWS Professional Services, Amazon Web Services & Greg Dumont - Director Technology, Nielsen
Customers looking to migrate to AWS often ask the following questions at the beginning of their journey “What is the right AWS account structure and VPC design for me?” and “How do I minimize the impact to my IT operations?” This session discusses the different account structures, VPC design patterns and network deployment architectures to align to different Enterprise IT Operating Models. We will also discuss the implications each pattern has with respect to Security, Finance and Operations. Nielsen, a global enterprise providing an understanding of what consumers watch and what consumers buy, will discuss the decision making process for their AWS account and VPC design. Nielsen will also discuss the implications of their decision. This session will benefit IT architects, managers and technology strategists to:
Kevin Aylward - Director, AWS Global Center of Excellence, Computer Sciences Corporation & Josh Levine - Global Alliance Executive, AWS, Computer Sciences Corporation
As you begin to move out of your data center and develop a cloud-first strategy, you'll need support for large-scale migrations to AWS. In this session, CSC shares details about the journey to AWS by some of our largest enterprise customers. We provide best practices for planning your large-scale migrations and focus on business processes in addition to technology. We show how CSC used this approach to migrate to AWS as part of our separation last year into two publicly traded companies: CSC and CSRA. In less than six months, CSC took our 56-year-old company and broke it into two companies, one of which was brand new and without any infrastructure or enterprise applications. We explain how we leveraged the AWS Partner ecosystem to achieve this incredible IT challenge.
Ryan Kennedy - Principal Automation Architect, 2nd Watch, Inc. & Christopher Nolan - Director of Product, 2nd Watch & Chris Resch - EVP, Cloud Solutions, Sales, 2nd Watch
Every enterprise knows by now that it can save money by simply lifting and shifting workloads to the cloud, but many are missing the larger opportunity to also make money by moving. While quick costs savings are good for the bottom line, they do little to move the top line numbers. To achieve both savings and earnings, corporate thinking about technologies must change in order to enable faster processes leveraged enterprise-wide. In this session we will explore multiple customer success stories where the customers have evolved from leveraging basic compute and storage products (EC2 and S3) to integrating new services into operations by leveraging Lambda, DynamoDB, CodeDeploy, etc. Once this is achieved, enterprises are enabled to manage and deploy code rapidly in a programmatic and elastic secure network, ensuring governance and security standards across the globe. We will look at the migration process trusted by hundreds of clients as well as how to cope with the process and people components that are so important to enable agility, while focusing heavily on the technology. The Coca-Cola Company (TCCC) and 2nd Watch story will dive deep into the technology that allows TCCC to manage hundreds of AWS Accounts, hundreds of workloads, thousands of instances, and hundreds of business partners around the globe. TCCC’s Configuration Management System has Puppet at the core and relies on over a dozen core and emerging AWS products across accounts, availability zones and regions. This complex and globally-available system ensures all of TCCC’s workloads in AWS meet corporate policies but also allows for rapid scale of both consumer and enterprise workloads.
Eric Tachibana - Area Practice Manager for APAC, China, and Japan, Amazon Web Services
Incumbent enterprises face dramatically competitive landscapes, with threats from almost every direction. Protecting your core business and innovating for the future is a delicate balancing act. Innovating as fast as a startup becomes a core competency, but failed new product innovation wastes time and resources. In this session, IT managers and professionals learn how running a lean enterprise can be a powerful framework for leading enterprise-scale innovation as effectively and fast as a startup.
Keith Jarrett - Business Development Manager - Cloud Economics, Amazon Web Services
Organizations where cloud adoption has matured into broader enterprise deployment are facing the need to better manage and control their costs and expenditures. Cost optimization at scale is a process that involves a number of changes across the business, including technical, organizational and cultural transformation. In this session, you will learn the fundamentals of cost optimization and how this can be used to help your organization drive costs down and still being able to meet capacity, demand and organizational requirements. Key topics being discussed are right sizing services, optimizing purchase models and implementing a culture of cost management.
Marc Trimuschat - Head of Storage Technical Business Development, Amazon Web Services
When evaluating and planning migrating your data from on premises to the Cloud, you might encounter physical limitations. Amazon offers a suite of tools to help you surmount these limitations by moving data using networks, roads, and technology partners. In this session, we discuss how to move large amounts of data into and out of the Cloud in batches, increments, and streams.
Steven Nelson - Global Storage Segment Technical Lead, Amazon Web Services & Isaiah Weiner - Sr. Mgr. Solutions Architecture, Amazon Web Services
This sessions is for architects and storage admins seeking simple and non-disruptive ways to adopt cloud platforms in their organizations. You will learn how to deliver lower costs and greater scale with nearly seamless integration into your existing Backup ad Recovery processes to achieve fast, simple wins that demonstrate the scale and flexibility of cloud services for storage. Services mentioned: S3, Glacier, Snowball, 3rd party partners, Storage Gateway, and cloud data migration services.
Mario Thomas - Senior Consultant, Amazon Web Services & Greg Cope - Head of Platform Architecture and Security, Financial Times Ltd
Many organizations rely on a TCO calculation to build a business case for migration to the Cloud. Unfortunately, this approach overlooks organizational readiness and wider economic factors, such as the total cost and the benefits of migration. This session provides executives with a toolkit for measuring readiness and developing a compelling business case.
Amit Sharma - Partner Solutions Architect, Amazon Web Services & Craig Bruce - Scientific Software Development Manager, OpenEye Scientific & Jharrod LaFon - Chief Cloud Engineer, OpenEye Scientific & Anthony Nicholls - CEO, OpenEye Scientific
Fortune 500 companies are increasingly using cloud services to run enterprise workloads to improve security, increase agility, and enable scale. Learn how OpenEye is running their AWS-native platform and workflow engine to support collaboration and data sharing at large pharmaceutical companies like Pfizer. In this session, OpenEye will share cloud best practiced around security controls, cross-departmental collaboration across the enterprise, and agility at scale. Attendees will gain practical tips for using AWS in the enterprise and healthcare industries.
Srinivas Margashayam - Sr Solutions Architect, Amazon Web Services & Derek Masseth - Deputy Chief Information Officer, Chief Technology Officer, Ofc of the Chief Info Officer, University of Arizona & Lauren Babson - Director, Enterprise Technical Services, University of Arizona
As enterprises seek to fully realize the scalability, availability, security, and agility of cloud computing, many are learning that they must migrate core business applications, such finance and human resource systems, not just public-facing websites. The University of Arizona shares how they addressed the challenge of creating a more efficient and agile IT operation. After initial success in migrating their Research Administration system, UA became the first major university to migrate their PeopleSoft application environments—Learning (EL), Human Resources (HR), and Student (SA)—to AWS. Learn about the university's enterprise migration roadmap, lessons learned, plans and budgets, and the tools available to support major enterprise migrations. See how to create a roadmap to cloud computing, transform application development, deploy mission-critical workloads at scale, and retrain staff.
Arjun Shah - Senior Manager and Practice Director, Accenture & Steve Hull - Director, Cloud Strategy and Transformation, T-Mobile, USA & Miha Kralj - Managing Director
After launching several thousand EC2 instances in the cloud, we've learned that the key to running an IT fleet successfully on AWS is enforcing operational and economic discipline. As AWS service consumption grows, operational costs and overhead shouldn't grow linearly. Instead, IT should encourage new tenants that migrate from data centers to AWS to slowly shift toward a self-service delivery model and adopt the DevOps operations model. Creating and offering an AWS Cloud operations service catalog enables organizations to efficiently take full advantage of AWS' flexibility and modularity. T-Mobile, whose journey to AWS Cloud management started more than 2 years ago, uses a service catalog to enforce operational discipline in the Cloud. Their catalog is custom crafted for each Cloud-based IT workload. This session provides insight into the AWS Cloud operations strategy and its transformation, the creation of a Cloud operations service catalog, and how this approach supports reliable engineering on AWS. S
Lou De La Torre - Solutions Architect, Amazon Web Services
In this workshop, we will explore the different approaches to migrating Microsoft applications to AWS. We’ll walk through the concerns and considerations to take into account while planning a migration, and learn how to develop and implement a migration plan to move applications from on-premises (or traditional hosting) to AWS. This session will use a case study format to dive deep into the details of how to successfully plan an application migration. To keep it real, teams will work through planning a SharePoint migration that integrates in with an existing Active Directory.
Trevor Dyck - Sr. Product Manager, Messaging, AWS & Quinn Easterbrook - Director - Platform Engineering, Capital One
Enterprises rely on messaging to integrate services and applications and to exchange information critical to running their business. However, managing and operating dedicated message-oriented middleware and underlying infrastructure creates costly overhead and can compromise reliability. In this session, enterprise architects and developers learn how to improve scalability, availability, and operational efficiency by migrating on-premises messaging middleware to a managed cloud service using Amazon SQS. Hear how Capital One is using SQS to migrate several core banking applications to the cloud to ensure high availability and cost efficiency. We also share some exciting new SQS features that allow even more workloads to take advantage of the cloud.
Nirav Kothari - Principal Consultant - AWS Professional Services, Amazon Web Services & John Hutchinson - IT Director, Technip
Data center migrations are often long and complex projects that involve thousands of servers. Without automated tools, these migrations can consume hundreds of hours. The recently launched AWS Server Migration Service speeds and simplifies the migration of on-premises servers to the AWS Cloud. SMS automates incremental replication of live server volumes, speeding up the migration process and significantly reducing the manual tasks required for server migration. This session for IT managers introduces key features and benefits of SMS, and includes a demo of the SMS end-to-end workflow.
Ryan Hughes - Cloud Executive Advisor, Amazon Web Services & Duane Smith - Sr. Enterprise Architect, Cardinal Health
For many large enterprises, gathering data on your applications and developing a detailed business case for an all-in migration to AWS can be a time consuming activity. Instead of waiting on a full analysis to be completed, a global healthcare company decided to take a shortcut by reinventing the portfolio assessment process in a brilliantly simple manner. This new streamlined approach enabled them to immediately take advantage of the increased speed and innovative capabilities of AWS. In this session, IT directors and enterprise architects can learn these techniques to quickly identify which of their own applications fit cloud-friendly patterns; including, how to build a prioritized and multi-threaded migration queue that maximizes velocity and business value.
Brent Eubanks - VP Technology Optimization, Ticketmaster & Vittaly Tavor - CoFounder & VP Products, Cloudyn
Ticketmaster will share their playbill to managing and optimizing their AWS deployment in real-time. Learn how this ticket-sales giant was able to deploy, train and obtain utilization of the Cloudyn solution for their significant AWS deployment - and get rapid results.
ENT306: Transforming BP’s Cloud Environment at Industrial Scale
Paul Schuster - Director, Shared Services Archiecture, BP plc & Adam Temple - Cloud Architect, Cloudreach Europe Limited & David Ninnis - Senior Enterprise Architect, Cloud, BP
In this talk, we will describe the journey and approach to moving BP to the Cloud. The IT team at BP was faced with a huge legacy estate, a highly conservative risk posture, and the most complex and heavily outsourced environment of any of the Oil & Gas super-majors. Working with Cloudreach and AWS, we have been on a three-year change journey to design a global Cloud migration program. We will describe how our strategy has evolved, our 3rd generation approach and operating model for brokering Cloud services at scale, creating a controls framework for Cloud security, developing the business case to transition from a large fixed cost base and, finally, developing an at-scale migration approach. The scale and complexity of our business has thrown up many difficult operational issues and we hope that this talk will provide valuable insight, to other organisations on a similar journey, into how BP has tackled these challenges. This session is sponsored by Cloudreach.
J.R. Storment - Co-Founder & Chief Customer Officer, Cloudability
As your Reserved Instance portfolio grows, tracking your ROI becomes more critical. You’ll need the right tools, metrics, and strategies in place to make sure you’re maximizing RI savings and minimizing unused RI hours.
Shahbaz Alam - Manager, AWS Professional Services, Amazon Web Services & Rehan Qureshi - Canada Practice Manager, Professional Services, Amazon Web Services
Learn how to create a compelling business case for a large-scale migration to AWS. We present a framework and tools for creating your business case, and guidelines for using AWS services to maximize value and optimize cost for migrations to the AWS Cloud. Learn a new way of thinking about cost that includes automation, new technologies, organizational change, and other factors.
Allen Brown - Mgr, Solution Architects, Amazon Web Services
You have decided to shut down a data center and move your applications to AWS, and you know you’ll need some help. Whether you need just tools, boots on the ground to move those applications, or someone to teach you how, you can find help among the group of elite partners that the AWS Partner Network has assembled to assist customers with meeting their business objectives. In this session, you will discover the variety of services our partners provide, learn how to select the appropriate partner for your needs, and hear about the experiences of customers who used partners to accomplish big objectives on AWS.
Nirav Kothari - Principal Consultant - AWS Professional Services, Amazon Web Services & Jeffrey Gelb - Head of Foundational Engineering, Scholastic Inc
Learn about the 6 Rs of migrating, migration patterns, and how to map your portfolio to the patterns. Discover how your business drivers and overall IT strategy determine the migration patterns you need, and learn best practices for selecting the appropriate Rs.
Mandus Momberg - Partner Solutions Architect, Amazon Web Services
Learn about some of the most useful and popular tools that you can leverage at various stages of a migration project. These tools will allow your teams to focus on coordinating the migration and automating as many migration activities as possible.
Mark Ryland - Director, Solutions Architecture, Amazon Web Services & Tigran Khrimian - Vice President, Data Platforms, FINRA
Large-scale enterprise migration can be a complex undertaking, especially for organizations that re-architect solutions to leverage the benefits of the Cloud. FINRA, which regulates US equities and options markets, recently completed a 2.5-year migration and re-architecture of its Big Data platform. Their platform consumes billions of market events every day. FINRA has developed scalable platforms and services on AWS that enable migrating enterprise applications and business functions to the Cloud quickly. Their data management platform takes advantage of AWS storage and compute products. In this session, IT influencers and decision makers will learn lessons from FINRA’s migration, including how to create an enterprise-class Cloud architecture and which technology skills are required for transitioning to the Cloud. We also share examples of the business value FINRA has realized.
Manish Mohite - Enterprise Solutions Architect, Amazon Web Services & Tejash patel - Chief Architect, Guardian Life Insurance
Migrating regulatory workloads can be complex and time consuming. Come learn how Guardian Life Insurance migrated their regulated workloads to the AWS Cloud. Guardian Life's hybrid cloud journey started with moving web and application tiers into AWS, while keeping legacy repositories on premises. Key planning activities included vendor selection, migration planning, hybrid architecture design, compliance ensurance, and AWS and partner tools selection. Guardian Life shares how they defined security, compliance, and data management controls and solutions, which enabled them to migrate entire applications to AWS. The session also covers lessons learned and key considerations for hybrid cloud architectures.
Keith Jarrett - Business Development Manager - Cloud Economics, Amazon Web Services & Elvy Lee - Senior Finance Leader, Intuit & Dieter Matzion - Staff Business Systems Analyst, Intuit Inc.
How do you optimize costs when migrating user-facing services from a proprietary data center to the Cloud? When Intuit faced this thorny question, they formed a cost-optimization team, built visibility into cost and usage data, decided when to use automation, and developed a culture of cost consciousness. Who do you select for this cross-functional team? How do you empower the team to reach across business units? Which metrics should you focus on first? How do you align your organization with this cause? Which roadblocks and issues could you face along the way? How will you create an internal chargeback model? Hear from the Intuit team about their journey, learn from the challenges that they overcame, and hear about the results they've achieved.
Paul Bockelman - Sr. Solutions Architect, WWPS, Amazon Web Services & Matt Dreyer - Product Management, VMware
brings VMware’s enterprise class Software-Defined Data Center software to Amazon’s public cloud, delivered as an on-demand, elastically scalable, cloud-based VMware sold, operated and supported service for any application and optimized for next-generation, elastic, bare metal AWS infrastructure. This solution enables customers to use a common set of software and tools to manage both their AWS-based and on-premises vSphere resources consistently. Further virtual machines in this environment have seamless access to the broad range of AWS services as well. This session will introduce this exciting new service and examine some of the use cases and benefits of the service. The session will also include a VMware Tech Preview that demonstrates standing up a complete SDDC cluster on AWS and various operations using standard tools like vCenter.
Alan Garver - Professional Services Consultant, Amazon Web Services & Jamie Greco - SVP Technical Program Management, Citi & Chuck Dudley - Director, Financial Services Engagements, Stelligent Systems LLC
Continuous delivery can be challenging, especially for enterprises that deal with strict compliance requirements, like those in the financial services sector. AWS and Stelligent frequently work together with many large financial services enterprises to build solutions that enable customers to run their business faster and more safely on AWS. Together, we help customers ensure the security of the source code used to trigger builds, insert strict business controls at run time, and continuously inspect running infrastructure to ensure compliance. In this session, we share highly effective techniques that you can incorporate into your continuous delivery system to provide bank-level controls and security, and faster deployments. We explore a strong encryption pattern for handling build artifacts in a continuous delivery pipeline, a simple process for inspecting AWS CloudFormation templates to ensure that business rules are in compliance before a template makes AWS API calls, and a runtime inspector that uses AWS Lambda and AWS Config rules to ensure that running infrastructure is always in compliance.
Chris Johnson - Solutions Architect & Michael Heimlich - Solution Delivery Manager, AstraZeneca & Ronen Artzi - RDI Enterprise Architect Science and Cloud Evangelist, AstraZeneca
AstraZeneca is a global science-led biopharmaceutical company developing innovative medicines used by millions of patients worldwide. With AWS, AstraZeneca processed more exomes in 20 days than during the previous 3 years, enabling scientists to receive results more quickly, develop medicines faster, and treat more patients sooner. AstraZeneca was able to identify ~20% more patients with actionable variants of major cancer types by using VarDict, an internally developed, open source contributed variant caller, at scale in the Cloud. Learn how AstraZeneca used the AWS Cloud Services to rapidly develop and scale an asynchronous architecture to meet an urgent business opportunity, accelerating the speed of scientific discovery in a cost effective manner.
Milind Halapeth - Vice President, Wipro Technologies & David Watterson - Head of IT, Broadspectrum
Join Broadspectrum as they share how they achieve their business goals using a cloud-first IT strategy and AWS for 'as a Service' deployments. To support new customer projects, Broadspectrum frequently needs to set up new sites or offices. This often requires setting up infrastructure for a specific site for only the duration of the project. Learn how Broadspectrum leverages AWS and Wipro's Boundary Less Data Center Solution to enable on-demand provisioning of 'site-in-a-box'. Gard Little, analyst from IDC, Stephen Orban, AWS Head of Enterprise Strategy, and Ramesh Nagarajan, SVP of Integrated Services at Wipro, join the discussion. Session sponsored by Wipro.
Saju Sankaran Kutty - Vice President, Infosys limited & Abhijit Shroff - Associate Vice President & Practice Head - Hybrid Cloud and Hybrid IT Autonomics Platforms, Infosys Limited
To keep pace with today’s hybrid IT landscape, enterprises have to innovate faster while ensuring compatibility with legacy IT systems. They also need to reduce costs and increase business value through responsive infrastructure backed by flexible, intelligent, and context-aware agile IT processes. To do this, they need service assurance in four distinct areas:
FIN: Financial Services
Pawan Agnihotri - Principal Solutions Architect -- Global Financial Services, Amazon Web Services
In this session, we provide programmatic guidance on building tools and applications to detect and manage fraud and unusual activity specific to financial services institutions. Payment fraud is an ongoing concern for merchants and credit card issuers alike and these activities impact all industries, but are specifically detrimental to Financial Services. We provide a step-by-step walkthrough of a reference solution to detect and address credit card fraud in real time by using Apache Apex and Amazon Machine Learning capabilities. We also outline different resource and performance optimization options and how to work data security into the fraud detection workflow.
Felix Candelario - Global Accounts Solutions Architect for Financial Services, Amazon Web Services & Benjamin Feldon - Enterprise Solutions Architect, Amazon Web Services
Modern financial services organizations rely heavily on technology and automated systems to run business-as-usual. However, if this technology were interrupted by natural disasters or other events, there could be a devastating impact on investors and market participants, and in turn your reputational brand. In this session, we provide a step-by-step disaster recovery solution employed by a major exchange. This solution leverages Amazon EC2 Container Service to provide Docker containers, Weave Net to support a multicast overlay network that enables high volume multicast feeds in a cloud environment, and AWS CloudFormation for the ability to easily create and manage AWS assets. The session also covers the importance of redundancy (not just operationally, but for SEC compliance reasons as well) and how financial services organizations can increase geographical diversification of their primary and disaster recovery data centers. We dive deep into each major component of the solution.
Felix Candelario - Global Accounts Solutions Architect for Financial Services, Amazon Web Services & Alan Garver - Professional Services Consultant, Amazon Web Services & Jamie Greco - SVP Technical Program Management, Citi & Chuck Dudley - Director, Financial Services Engagements, Stelligent Systems LLC
Continuous Delivery can be challenging, especially for enterprises that deal with strict compliance requirements like those in the financial services sector. AWS and Stelligent frequently work together with many large Financial Services Enterprises to help incorporate their capabilities securely on the Cloud. From security of the source code used to trigger builds, to the insertion of strict business controls at run time, and out to the continuous inspection of the running infrastructure to ensure compliance, we are helping to build capabilities that are enabling them to run their business faster and safer on AWS.
Pawan Agnihotri - Principal Solutions Architect -- Global Financial Services, Amazon Web Services & Francis Aherne - Global Head of SABRE Development, Standard Chartered Bank
In this session, we explain how Financial Services organizations can leverage AWS grid computing capabilities to perform large-scale calculations for risk management purposes. Numerous financial services companies face the same basic challenge: modeling multiple scenarios with different risk factors simultaneously or in quick succession in order to make informed decisions that maximize gains and minimize financial loss. Some examples of these workloads include Monte Carlo simulations, price model validation and back-testing, and risk calculations for hedging and capital optimization strategies. We provide programmatic guidance around what AWS services to use when running a grid computing cluster that requires thousands of cores and specific industry use cases and key benefits around speed and costs that the AWS platform, auto-scaling capabilities, and various compute services can help achieve.
Paul Durbin-Johnson - Director, Technical Operations, Electronic Arts & Bill Weiner - SVP Operations, 47Lining & Mark McBride - Senior Software Engineer, Electronic Arts - Capital Games
In November 2015, Capital Games launched a mobile game accompanying a major feature film release. The back end of the game is hosted in AWS and uses big data services like Amazon Kinesis, Amazon EC2, Amazon S3, Amazon Redshift, and AWS Data Pipeline. Capital Games will describe some of their challenges on their initial setup and usage of Amazon Redshift and Amazon EMR. They will then go over their engagement with AWS Partner 47lining and talk about specific best practices regarding solution architecture, data transformation pipelines, and system maintenance using AWS big data services. Attendees of this session should expect a candid view of the process to implementing a big data solution. From problem statement identification to visualizing data, with an in-depth look at the technical challenges and hurdles along the way.
David Green - Enterprise Solutions Architect, Amazon Web Services & Dustin Pham - Principal Engineer, Sony Interactive Entertainment & Alexander Filipchik - Principal Software Engineer, Sony Interactive entertainment
As systems and user bases grow, a once abundant resource can become scarce. While scaling out PlayStation services to millions of users at over a 100,000 requests/second, network throughput became a precious resource to optimize for. Alex and Dustin talk about how the microservices that power Playstation achieved low latency interactions while conserving on precious network bandwidth. These services powered by Amazon Elastic Load Balancing and Amazon DynamoDB benefitted from soft-state optimizations, a pattern that is used in complex interactions such as searching through a user’s social graph in sub 100 ms, or a user’s game library in 7 ms. As a developer utilizing Amazon Web services, you will discover new patterns and implementations which will better utilize your network, instances, and load balancers in order to deliver personalized experiences to millions of users while saving costs.
Kevin Ashman - Technical Evangelist, Amazon Web Services & Daniel Winters
This session focuses on the cloud-connected features of Amazon Lumberyard. We introduce Amazon GameLift core concepts, complete a step-by-step walkthrough of deploying a multiplayer game to the cloud, and demonstrate how to rapidly scale a game based on player demand. We also cover Amazon Lumberyard’s Cloud Canvas, which empowers developers to leverage services such as databases, data storage, and methods to execute game logic within AWS without needing to configure or maintain servers. Finally we discuss integrating Twitch to your Lumberyard game with Twitch ChatPlay and JoinIn.
Adam Rozumek - Systems Engineer, Riot Games
This session covers how the data team at Riot Games utilizes ECS to consolidate and improve disparate deployment and hosting strategies across a wide range of applications deployed via Docker containers. The team will share how cluster management and container orchestration through ECS enables the team to quickly adopt and evolve shared service hosting solutions as Riot continues its journey towards becoming a multi-game studio.
Ujjwal Ratan - Solutions Architect, Amazon Web Services & Rich Uhl - CTO / Founder, 1Strategy & Arnoud Otte - Assistant Director Cloud and Data Architecture, Cambia Health Solution
Working with Amazon Web Services “AWS” and 1Strat
Nathan Beyer - Senior Director & Distinguished Engineer, Cerner Corporation
This talk will feature Cerner’s
Scott Paddock - Security Solutions Architect, Amazon Web Services & Matt Ferrari - CTO, ClearDATA
This session will demonstrate how to embrace DevSecOps to improve your security and compliance agility and posture within the highly regulated HIPAA environment. We will cover compliance frameworks, data decoupling strategies to fully utilize AWS, and best practices learned from the industry most active cloud adopters.
Chris McCurdy - Solutions Architect Specialist in Healthcare and Lifesciences, Amazon Web Services & Nitin Gujral - Director of Software Development, Boston Children's Hospital
Alexa, what is the Internet of Things? Now that technology is small enough to be embedded in everyday devices, Healthcare has an opportunity to exploit the extraordinary potential of connecting ordinary devices. In this presentation, we explain how to rapidly build an IoT system and how to drive the Cloud with your voice on an Amazon Echo. In addition to describing how to use Alexa, we explore using AWS IoT, Lambda, Amazon SNS, and DynamoDB.
IOT: IoT Mini-Conference
Jon Turow - Principal Product Manager, Amazon Web Services & Bart Vercammen - VP Product Management, Technicolor & Emilie Megel , Technicolor
AWS has launched AWS Greengrass, a platform that extends the AWS Cloud onto your devices so they can act locally on the data they generate, while still taking advantage of the cloud. In this session we will talk about how Greengrass works and what you can do with it. You will also hear from early customers who will discuss their use cases for Greengrass and how it fits into their overall IoT strategy.
Satyen Yadav - General Manager, Internet of Things, Amazon Web Services & Art Miller - Senior Director, Smart Cities / Commercial and Industrial IoT, Qualcomm Technologies Inc. & Andy Daudelin - VP - Cloud Solutions, AT&T & Fabio Veronese - Head of Infrastructure and Technological Services, Enel & Rich Finn - IoT Evangelist, AT&T
AWS IoT edge and device services make it easy to get started and scale quickly along with your business needs. Medical equipment, industrial machinery, building automation, and simple device to trigger services, are just a few physical-world use cases that are benefiting from elastic cloud computing while meeting the local execution requirements and real time responsiveness. This session covers the intersection between the device and cloud industries, and the way AWS and our customers will shape the future of those industries together. We will showcase how our customers are using AWS IoT Button, the IoT Device SDKs, and other AWS services to improve the existing business models, invent new way of working, and balance the benefits of the cloud services with the need for local execution.
Shyam Krishnamoorthy - Senior Manager, SW Development, Amazon Web Services & Tor Boyesen - Associate Product Manager, HomeAdvisor & John Humphreys - VP Product Management, HomeAdvisor
Enterprises can use the AWS IoT button to provide unique innovative experiences to their consumers, simplifying the customer interface, and increasing engagement and brand loyalty. Enterprises and developers alike can use this single-purpose device to build fully customized IoT applications, or select from a list of predefined “blueprints” to instantly configure the device for unique use cases.
Antony Passemard - Head of Product Management - AWS IoT Platform, Amazon Web Services & Mike Gartner - Senior Solutions Architect, Amway & Everette Binger - SR. Solution Architect, Amway
What’s new with AWS IoT? This is an Introduction to the AWS IoT Platform and an overview of new features. Join us for a discussion on the features launched over the last year, and the best practices on how to use the AWS IoT Platform to get your device data into the cloud.
John Rotach - Software Development Engineer, Amazon Web Services & Ryan Burke - Software Development Manager, Amazon Web Services
Establishing connectivity between devices and the cloud is the first step to building great IoT applications. Generating insights with these data streams, and acting on it in real-time, that creates value for your business. Join us and learn how to expand the operational picture of your IoT solution with AWS IoT. We will show you how to use metrics to enable new possibilities in generating insights and decision making engines. And we will connect to Salesforce to leverage your CRM data and empower service agents in this IoT context, using a solar panel monitoring and maintenance as a demonstration of AWS IoT features paired with Salesforce IoT Cloud.
Anton Shmagin - Cloud Security Architect, Sr. Consultant, Amazon Web Services & Jan Metzner - Specialist Solutions Architect, IoT, Amazon Web Services & Kerry Maletsky - Senior Director Security Architecture, Microchip Technology
Only year ago we launched AWS IoT, and at re:Invent we showed how AWS IoT makes it easy to secure millions of connected devices. However, we have learned from our customers that a number of unique security challenges for the Internet of Things (IoT) exist.
Sarah Cooper - GM IoT Solutions, Amazon Web Services & OLGA LAGUNOVA - Chief Data and Analytics Officer, Pitney Bowes
In 2004, approximately 400 billion fax pages were sent. Twelve years later, it’s about 4% of that number. The pace of technological change is rapid, but most devices live in the field for 10 to 15 years. It’s hard to maintain competitive value in the face of constant technology improvement, but IoT is changing that. We’ll examine the architectures that allows AWS IoT customers like Pitney Bowes to connect devices to the cloud and enrich the client experience though personalized analytics and recommendations, automated supplies replenishment, and just-in-time self-service.
Michael Garcia - Technical Program Manager, AWS IoT, Amazon Web Services & Tom Soderstrom - IT Chief Technology and Innovation Officer, JPL
Jet Propulsion Laboratory is a well-known innovator in outer space, particularly in its search for 'life out there'. JPL is now innovating in the physical space to improve 'life here'. AWS IoT is critical to their innovations. See a re:Invent preview about how JPL, as an early adopter of AWS IoT, has prototyped voice control to ask questions of the room, the budget, or the system. They've also used it for controlling lights and sound to detect cyber security threats, rapid prototyping of robots, low-cost virtual windows to the outside, and much more. The results have been excellent. JPL will demonstrate and talk about these prototypes, including what worked and what didn’t. They will also share the promise integrated serverless computing holds.
IOT306: IoT Visualizations and Analytics
Asha Chakrabarty - Senior Solutions Architect, Amazon Web Services
In this workshop, we focus on visualizations of IoT data using ELK, Amazon Elasticsearch, Logstash, and Kibana or Amazon Kinesis. We will dive into how these visualizations can give you new capabilites and understanding when interacting with your device data from the context they provide on the world around them.
Dirk Didascalou - VP AWS IoT, Amazon Web Services & Brian Pokorny - Sr. Manager, Media and Digital, MillerCoors & Mark Holdsworth - Chief Advisor, Integrated Operations, Growth & Innovation, Rio Tinto & Gary Gutknecht - SVP of Connected Home Division, Technicolor & Tom Soderstrom - IT Chief Technology and Innovation Officer, JPL & Jonathon Ballon - VP | GM of IoT, Intel & Dale Wiggins - VP & GM HealthSuite Digital Platform, Philips
In this general session, AWS IoT experts will present an in-depth look at the current state of the Internet of Things. Learn about trends and industry use cases. Hear how other organizations are using AWS IoT to connect devices to the cloud. Explore some of the most recent IoT announcements as we kick off the IoT re:Source Mini Con.
IOT308: Smart Home Solution with Amazon Alexa
Robert Mueller - IoT Solutions Architect, Intel IoT Group, Intel & Lakshmi Talluru - Director, IOT Solutions, Intel
Simply connecting the 'things' that were never connected before as part of the Internet-of-Things is leading to new data insights that translate into meaningful change. AWS and Intel are working together to provide a secure, scalable edge-to-cloud solution for IoT applications. Intel gateways utilize Windriver Helix Device Cloud to authenticate the device with AWS IoT and initiate secure data transport as well as provide a framework for active edge device management and over the air software and security updates. With AWS and Intel, you can implement an IoT solution quickly and with minimal upfront investments, seamless connectivity, and deliver enhanced security from device to network to cloud, then use AWS Big Data services to drive business insight.
Marc Brooker - Principal Engineer, Amazon Web Services
You can now execute Lambda’s almost anywhere – originating in the cloud, and on connected devices with AWS Greengrass. This advanced technical session explores Lambda Functions and what it means to use them across these diverse environments. We will treat the cloud as the ‘brain’, using local Lambda’s for local executions. This way devices can react instinctively, much like the autonomic nervous system, operating in the periphery and responsible for collecting and filtering information, implementing simple and time-sensitive local actions reflexively.
Olawale Oladehin - Solutions Architect, Amazon Web Services & Ben Kehoe - Cloud Robotics Research Scientist, iRobot
This session will cover common customer implementations and patterns for building connected/smart home implementations with AWS IoT. This includes the end-user experience for onboarding a smart home appliance and then integrating it with the AWS ecosystem (for targeted push notifications, predictive maintenance, and so on). iRobot will join us to discuss their smart home integrations with the Roomba 980 and AWS IoT.
Brett Francis - Principal Solutions Architect, Specialist, Amazon Web Services & Asif Khan - Solutions Architect, Amazon Web Services & John Crawford - CTO, ndustrial.io & Jason Massey - Founder and CEO, ndustrial.io
30% of global food produce is wasted in the supply chain: storage, movement, and delivery. By using AWS IOT to enable sensors to manage the supply chain and big data to understand patterns, industrial companies can gain efficiencies in electricity and transportation.
Olivier Klein - Emerging Technologies Solutions Architect, Asia-Pacific, Amazon Web Services & Markku Lepisto - Principal Technology Evangelist, Amazon Web Services
Experience how live virtual 3D worlds rendered with Amazon Lumberyard, a free, cross-platform, 3D game engine, interconnect with IoT devices in the real world. This session will illustrate how AWS IoT can be used to remotely control animate objects with Bluetooth, such as a Sphero robot. By using AWS Lambda and AWS IoT, we will create a bi-directional communication between moving robots that detect collisions and a virtual world rendered in Amazon's game engine. By using AWS IoT with the Alexa Skills Kit and the Amazon Echo, we will learn how to control the physical and virtual robots through voice.
LFS: Life Sciences
Patrick Combes , Amazon Web Services & Ryan Ulaszek - Software Engineer, Human Longevity & Lance Smith - IT Director, Celgene & Stephen Terrell - Senior Cloud Engineer, Human Longevity Inc.
This session discusses the architecture, formation, and usage of a collaborative HPC/big data scientific research and analysis environment on AWS. The pharmaceutical industry trend toward joint ventures and collaborations has created a need for new platforms in which to work together. We'll dive into architectural decisions for building collaborative systems. Examples include how such a platform allowed Human Longevity, Inc. to accelerate software deployment to production in a fast-paced research environment, and how Celgene uses AWS for research collaboration with outside universities and foundations.
Scott Paddock - Security Solutions Architect, Amazon Web Services & Jeff Feist - Cloud Architect, Merck & Co., Inc.
Life sciences organizations running regulated workloads in the cloud can move from point-in-space testing of their environment to near real-time testing to achieve continuous compliance with the mandates of auditors and regulation entities. Get deep insights into some of the AWS services used to accomplish continuous compliance such as Amazon CloudTrail, Amazon CloudWatch, AWS Config, Amazon VPC, Amazon S3, and Amazon EC2. Get real-world use cases of how heavily regulated environments within Merck maintain governance and control over a shared environment. We also discuss the automated tools used by Merck to eliminate manual processes and streamline IT management.
Dario Rivera - Solutions Architect, Amazon Web Services & Neeraj Verma - Solutions Architect, Amazon Web Services & Saurav Mahanti - Senior Manager, Information Systems, Amgen
For discovery-phase research, life sciences companies have to support infrastructure that processes millions to billions of transactions. The advent of a data lake to accomplish such a task is showing itself to be a stable and productive data platform pattern to meet the goal. We discuss how to build a data lake on AWS, using services and techniques such as AWS CloudFormation, Amazon EC2, Amazon S3, IAM, and AWS Lambda. We also review a reference architecture from Amgen that uses a data lake to aid in their Life Science Research.
Angel Pizarro - Technical Business Development Manager, Amazon Web Services & Brian O'Connor - Technical Director, Analysis Core, UCSC Genomics Institute
The PanCancer Analysis of Whole Genomes (PCAWG) project is a large-scale, highly distributed research collaboration designed to identify common patterns of mutations across 2,800 cancer genomes. The use of public and private clouds were instrumental in analyzing this dataset using current best practice containerized pipelines. This session describes the technical infrastructure built for the project, how we leveraged cloud environments to perform the “core” analysis, and the lessons learned along the way.
MAC: Machine Learning Mini-Conference
Ranju Das - Sr. Manager, Amazon Web Services & Peter Hallinan - Head, CV Data, A9.com
Jump-start your machine learning project by using the crowd to build your training set. Before you can train your machine learning algorithm, you need to take your raw inputs and label, annotate, or tag them to build your ground truth. Learn how to use the Amazon Mechanical Turk marketplace to perform these tasks. We share Amazon's best practices, developed while training our own machine learning algorithms, and walk you through quickly getting affordable and high-quality training data.
Nikko Strom - Sr. Principal Scientist & Arpit Gupta - Applied Scientist, AMAZON
Neural networks have a long and rich history in automatic speech recognition. In this talk, we present a brief primer on the origin of deep learning in spoken language, and then explore today’s world of Alexa. Alexa is the AWS service that understands spoken language and powers Amazon Echo. Alexa relies heavily on machine learning and deep neural networks for speech recognition, text-to-speech, language understanding, and more. We also discuss the Alexa Skills Kit, which lets any developer teach Alexa new skills.
Ranju Das - Sr. Manager, Amazon Web Services
This session will introduce you to Amazon Rekognition, a new service that makes it easy to add image analysis to your applications. With Rekognition, you can detect objects, scenes, and faces in images. You can also search and compare faces. Rekognition’s API lets you easily build powerful visual search and discovery into your applications. With Amazon Rekognition, you only pay for the images you analyze and the face metadata you store. There are no minimum fees and there are no upfront commitments.
Rafal Kuklinski - Senior Software Development Manager, Amazon Web Services & Manoj Sindhwani - Director, Software Development, Amazon Alexa & Stacy Adams - Head of Marketing, GoAnimate & Severin Hacker - CTO, Duolingo
This session will introduce you to Amazon Polly, a new deep learning service that turns text into lifelike speech. Polly enables existing applications to speak as a first class feature and creates the opportunity for entirely new categories of speech-enabled products – from mobile apps and cars, to devices and appliances. Polly includes 47 lifelike voices and support for 24 languages, so you can select the ideal voice and distribute your speech-enabled applications in many geographies. Polly is easy to use – you just send the text you want converted into speech to the Polly API, and Polly immediately returns the audio stream to your application so you can play it directly or store it in a standard audio file format, such as MP3. Polly supports Speech Synthesis Markup Language (SSML) tags like prosody so you can adjust the speech rate, pitch, or volume. Polly is a secure service that delivers all of these benefits at high scale and at low latency. You can cache and replay Polly’s generated speech at no additional cost. Polly lets you convert 5M characters per month for free during the first year. Polly’s pay-as-you-go pricing, low cost per request, and lack of restrictions on storage and reuse of voice output make it a cost-effective way to enable speech synthesis everywhere. Join this session to learn more and find out how you get can started with Amazon Polly, today!
Andres Rodriguez - Solutions Architect, Intel & Juan Carlos Riveiro - CEO, Vilynx
Deep learning continues to push the state of the art in domains such as video analytics, computer vision, and speech recognition. Deep networks are powered by amazing levels of representational power, feature learning, and abstraction. This approach comes at the cost of a significant increase in required compute power, which makes the AWS cloud an excellent environment for training. Innovators in this space are applying deep learning to a variety of applications. One such innovator, Vilynx, a startup based in Palo Alto, realized that the current pre-roll advertising-based models for mobile video weren’t returning publishers' desired levels of engagement. In this session, we explain the algorithmic challenges of scaling across multiple nodes, and what Intel is doing on AWS to overcome them. We describe the benefits of using AWS CloudFormation to set up a distributed training environment for deep networks. We also showcase Vilynx’s contributions to video discoverability, and explain how Vilynx uses AWS tools to understand video content. This session is sponsored by Intel.
Swaminathan Sivasubramanian - General Manager, Amazon Web Services & Salil Verma - Sr Director, IT, OhioHealth & Dan Law - Chief Data Scientist, Motorola Solutions
With the growing number of business cases for artificial intelligence (AI), machine learning (ML) and deep learning (DL) continue to drive the development of cutting edge technology solutions. We see this manifested in computer vision, predictive modeling, natural language understanding, and recommendation engines. During this full afternoon of sessions and workshops, learn how you can develop your own applications to leverage the benefits of these services. Join this State of the Union presentation to hear more about ML and DL at AWS and see how Motorola Solutions is leveraging these state-of-the-art technologies to solve public safety challenges, and how Ohio Health intends to inject AI into the medical system.
MAC301: Transforming Industrial Processes with Deep Learning
ernesto gonzalez - Sr. Manager, Applied Research & Marshall Tappen - Principal Applied Scientist, Computer Vision
Deep learning has revolutionized computer vision by significantly increasing the accuracy of recognition systems. This session will discuss how the Amazon Fulfillment Technologies Computer Vision Research team has harnessed deep learning to identify inventory defects in Amazon’s warehouses. Beginning with a brief overview of how orders on Amazon.com are fulfilled, the talk will describe a combination of hardware and software that uses computer vision and deep learning that visually examine bins of Amazon inventory to locate possible mismatches between the physical inventory and inventory records. With the growth of deep learning, the emphasis of new system design shifts from clever algorithms to innovative ways to harness available data.
Mick Bass - CEO, 47Lining & Daryan Dehghanpisheh - SVP, Digital Strategy, The Howard Hughes Corporation
The Howard Hughes Corporation partnered with 47Lining to develop a managed enterprise data lake based on Amazon S3. The purpose of the managed EDL is to fuse relevant on-premises and third-party data to enable Howard Hughes to answer its most valuable business questions. Their first analysis was a lead-scoring model that uses Amazon Machine Learning (Amazon ML) to predict propensity to purchase high-end real estate. The model is based on a combined set of public and private data sources, including all publicly recorded real estate transactions in the US for the past 35 years. By changing their business process for identifying and qualifying leads to use the results of data-driven analytics from their managed data lake in AWS, Howard Hughes increased the number of identified qualified leads in their pipeline by over 400% and reduced the acquisition cost per lead by more than 10 times. In this session, you will see a practical example of how to use Amazon ML to improve business results, how to architect a data lake with Amazon S3 that fuses on-premises, third-party, and public data sets, and how to train and run an Amazon ML model to attain predictive accuracy.
Jonathan Fritz - Sr. Product Manager, Amazon Web Services & Jasjeet Thind - Sr. Director, Data Science & Engineering, Zillow Group
Customers are adopting Apache Spark ‒ an open-source distributed processing framework ‒ on Amazon EMR for large-scale machine learning workloads, especially for applications that power customer segmentation and content recommendation. By leveraging Spark ML, a set of machine learning algorithms included with Spark, customers can quickly build and execute massively parallel machine learning jobs. Additionally, Spark applications can train models in streaming or batch contexts, and can access data from Amazon S3, Amazon Kinesis, Amazon Redshift, and other services. This session explains how to quickly and easily create scalable Spark clusters with Amazon EMR, build and share models using Apache Zeppelin and Jupyter notebooks, and use the Spark ML pipelines API to manage your training workflow. In addition, Jasjeet Thind, Senior Director of Data Science and Engineering at Zillow Group, will discuss his organization's development of personalization algorithms and platforms at scale using Spark on Amazon EMR.
Vikram Anbazhagan - Sr Manager -Deep learning, Amazon Web Services & Michael (Mik) Cox - Data Scientist, Jet Propulsion Laboratory & Tom Soderstrom - IT Chief Technology and Innovation Officer, JPL
Amazon Lex is a service for building conversational interfaces into any applications using voice and text. With Lex, the same deep learning engine that powers Amazon Alexa is now available to any developer, enabling you to build sophisticated, natural language chatbots into your new and existing applications. Amazon Lex provides the deep functionality and flexibility of natural language understanding (NLU) and automatic speech recognition (ASR) to allow you to build highly engaging user experiences with lifelike, conversational interactions. In this introductory session, find out how Lex provides deep functionality and flexibility to empower you to define entirely new categories of products that are made possible through conversational interfaces.
Leo Dirac - Principal Engineer, Amazon Web Services
For many companies, recommendation systems solve important machine learning problems. But as recommendation systems grow to millions of users and millions of items, they pose significant challenges when deployed at scale. The user-item matrix can have trillions of entries (or more), most of which are zero. To make common ML techniques practical, sparse data requires special techniques. Learn how to use MXNet to build neural network models for recommendation systems that can scale efficiently to large sparse datasets.
Denis Batalov - Solutions Architect, Amazon Web Services
In this session, we take a specific business problem—predicting Telco customer churn—and explore the practical aspects of building and evaluating an Amazon Machine Learning model. We explore considerations ranging from assigning a dollar value to applying the model using the relative cost of false positive and false negative errors. We discuss all aspects of putting Amazon ML to practical use, including how to build multiple models to choose from, put models into production, and update them. We also discuss using Amazon Redshift and Amazon S3 with Amazon ML.
Vikram Anbazhagan - Sr Manager -Deep learning, Amazon Web Services & Ranju Das - Sr. Manager, Amazon Web Services & Federico Lucignano - Software Development Manager, Amazon Web Services & Rafal Kuklinski - Senior Software Development Manager, Amazon Web Services & Piotr Lewalski - Sr Technical Program Manager, Amazon Web Services
Amazon AI services bring natural language understanding (NLU), automatic speech recognition (ASR), visual search and image recognition, text-to-speech (TTS), and machine learning (ML) technologies within reach of every developer. Amazon Lex make it easy to build sophisticated text and voice chatbots, powered by Alexa; Amazon Rekognition provides deep learning-based image recognition; and Amazon Polly turns text into lifelike speech. In this workshop, you'll get a chance to use each of the new deep learning services. We'll see you there!
MAC401: Scalable Deep Learning Using MXNet
Alexander Smola - Director, Amazon Web Services & Mu Li - Sr. Applied Scientist, Amazon Web Services
Deep learning continues to push the state of the art in domains such as computer vision, natural language understanding and recommendation engines. One of the key reasons for this progress is the availability of highly flexible and developer friendly deep learning frameworks. During this workshop, members of the Amazon Machine Learning team will provide a short background on Deep Learning focusing on relevant application domains and an introduction to using the powerful and scalable Deep Learning framework, MXNet. At the end of this tutorial you’ll gain hands on experience targeting a variety of applications including computer vision and recommendation engines as well as exposure to how to use preconfigured Deep Learning AMIs and CloudFormation Templates to help speed your development.
Apaar Sadhwani - Researcher, Stanford & Jason Su - Research Assistant, Stanford University
Diabetic retinopathy (DR) is the leading cause of blindness for people aged 20 to 64, and afflicts more than 120 million people worldwide. Fortunately, vigilant monitoring greatly improves the chance to preserve one’s eyesight. This work used deep learning to analyze images of the retina and fundus for automated diagnosis of DR on a grading scale from 0 (normal) to 4 (severe). We achieved substantial improvement in accuracy compared to traditional approaches and continued advances by using a small auxiliary dataset that provided low-effort, high-value supervision. Data for training and testing, provided by the 2015 Kaggle Data Science Competition with over 80,000 high resolution images (>4 megapixels), required Amazon EC2 scalability to provide the GPU hardware needed to train a convolutional network with over 2 million parameters. For the competition, we focused on accurately modeling the scoring system, penalizing bad mistakes more severely, and combatting the over-prevalence of grade-0 examples in the dataset. We explored ideas first at low resolution on low-cost single-GPU instances. After finding the best methodology, we showed it could be scaled to equivalent improvements at high resolution, using the more expensive quad-GPU instances more effectively. This prototype model placed 15 out of 650 teams across the world with a kappa score of 0.78. We’ve now advanced the model via a new architecture that integrates the prototype and a new network specialized in finding dot hemorrhages, critical to identifying early DR. By annotating a small set of 200 images for hemorrhages, the performance jumped to a kappa of 0.82. We believe strategies that employ a bit more supervision for more effective learning are pivotal for cracking deep learning’s greatest weakness: its voracious appetite for data.
MAE: Media & Entertainment
Dustin Encelewski - Director of Virtualization and Management Products, Amazon Web Services & Marcial Rivas - IT Manager, Cinepolis
In this session, we show how to seamlessly transition VOD, live, and other advanced media workflows from on-premises deployments to the cloud. Cinépolis will provide an overview of their transcoding solution on AWS and how they have seamlessly expanded the solution increasing their customer reach. We'll show real world examples of the API calls used to configure and control all elements of the workflow including compression and origination. And how standard AWS services can be media-optimized with Elemental Technologies to form a robust live solution.
Usman Shakeel - Principal Solution Architect, AWS & Michael Koetter - SVP-Media Technology and Development, Turner Broadcasting
As Turner continues to make the transition from a traditional broadcast organization to a consumer-centric, data-driven media company, we are being challenged to re-think our approach to content supply. There is a need to achieve new levels of agility, flexibility and scalability to meet the rapidly evolving demands of our top media brands - including TBS, TNT, Cartoon Network, Adult Swim and CNN. To that end, we are transitioning the infrastructure that acquires, processes and distributes media for consumer-facing systems to the cloud. At the core of this environment is our Supply Chain Management application. The SCM app provides business and technical process management via an HTML based UI framework, State Machine, Rules Engine, Cost Model, Forms Service. We took advantage of several AWS specific services, including Lambda, S3, Dynamo DB, SNS, Elastic, Cloud Formation and Code Commit. The entire system is instance-less with all application code running in either the browser or within Lambda's. To ease development and debugging we created a method to run all JS libraries in the browser, switching to Lambda when we deploy with Code Commit. Cloud media processing infrastructure is BEING created on demand via an integration with SDVI. The SDVI and SCM apps exchange events and data via SNS and S3.
Michael Raposa - Principal Consultant, AWS & Dave Duvall - SVP - Infrastructure & Support Services, Discovery Communications & Brinton Miller - Senior Vice President, Technology Strategy & Architecture, Discovery Communications
Media delivery requirements are continually changing, driven by accelerating mobile, tablet, smart TV, and set-top technology advances. Broadcasters need agile solutions to the changing media and entertainment landscape that don't require multiyear projects with large upfront investments. In this session, we walk through Discovery Communications' migration of its broadcast playout and channel origination to AWS. Discovery Communications is a leader in nonfiction media, reaching more than 3 billion cumulative viewers in 220 countries and territories. Traditionally, broadcast origination for content delivered to telecommunications companies, cable TV, and satellite has existed only in on-premises data centers. In this session, we walk through Discovery's migration of broadcast playout supporting hundreds of channels worldwide to AWS. We show how Discovery has not only reduced their TCO but also has improved their agility by launching new channels on demand. We also walk through how channel origination is being deployed in a secure, automated fashion, and with a level of high availability that exceeds what is possible in a traditional data center.
Konstantin Wilms - Principal Solutions Architect, Amazon Web Services & Richard Welsh - CEO, Sundog Media
The process of making a film is highly complex, and comprises of multiple workflows across story development, pre-production, production, post-production and final distribution. Given the size and amount of media and assets associated with each stage, high performance infrastructure is often essential to meeting deadlines.
Mohit Srivastava - Principal Product Manager, Amazon Web Services & Amit Patel - General Manager, Amazon Web Services & Ross McKegney - Director, Platform, Box
AWS provides a range of services and tools to help you create industry leading, cloud-enabled mobile apps that can securely scale to millions of users globally. Join Amit Patel, GM of AWS Mobile, to hear our vision for mobile apps and the cloud, industry trends, recent product launches, and success stories directly from our customers. We'll walk through and demo the AWS Mobile offerings for building compelling cloud-enabled mobile apps and for engaging your app users. You’ll learn how to use these offerings (serverless – API Gateway/Lambda, Cognito, and new services) to make it easy to develop both your iOS and Android frontend, as well as your mobile backend.
Christian Lam - Analytics Engineer, gopro & David Winters - Big Data Architect, GoPro & Gaurav Kumar - Product Lead, GoPro, GoPro
As GoPro expands into content networks and launches new products, new challenges have appeared. One of the most critical challenges facing GoPro during this period of rapid growth is their ability to make effective use of massive amounts of data. Every day, GoPro collects increasing amounts of data generated by internet connected consumer devices (smart cameras, smart drones), GoPro mobile apps, GoPro content networks, GoPro e-commerce sales, and social media. This data ranges from raw camera logs to refined and well-structured e-commerce datasets. In the past, it took GoPro months to understand new inbound data and determine how to transform or augment it for analysis. To streamline this process and bridge the gap between tech-savvy engineers and data-savvy analysts, GoPro is creating an analysis loop, which informs product usage trends and product insights. This analysis loop serves a large ecosystem of GoPro executives, product managers, engineers, data scientists, and business analysts through an integrated technology pipeline consisting of Apache Kafka, Apache Spark Streaming, Cloudera’s distribution of Hadoop, and Tableau’s Data Visualization Software as the end user analytical tool. Session sponsored by Tableau Software.
Kadir Rathnavelu - Senior Product Manager, Amazon Web Services & Devika Chawla - Director of Engineering, Netflix
Companies around the world are using Amazon Simple Email Service (Amazon SES) to send millions of emails to their customers every day, and scaling linearly, at cost. In this session, you learn how to use the scalable and reliable infrastructure of Amazon SES. In addition, Netflix talks about their advanced Messaging program, their challenges, how SES helped them with their goals, and how they architected their solution for global scale and deliverability.
Pawel Wojnarowicz - Sr. Product Manager, Amazon Web Services & Ilya Epshteyn - Solutions Architect, Amazon Web Services & Greg Sypolt - Senior Engineer, Gannett
With the widespread and accelerating adoption of mobile devices, delivering high quality mobile experiences became essential for success in today's marketplace In this session, learn how you can deliver top quality apps by using AWS Device Farm for testing and debugging on hundreds of unique Android and iOS devices in the cloud. A Device Farm customer, Gannett, describes their journey from a time-consuming manual activity to fast and repeatable automation activity when using Chef, Scalr, AWS, and Jenkins. The new cloud infrastructure allows Gannett to scale on-demand to build and test 100+ Android native apps on Amazon EC2 instances and Device Farm to speed up their continuous integration pipeline with efficiency.
Sri Elaprolu - Global Lead, Public Sector IoT, Amazon Web Services & Kristin Boorse - Sr. Product Manager, Thorn & Paul Ames - SVP, Products and Technology, SST, Inc. (ShotSpotter)
Join us for a compelling session to learn how technological innovation can save lives. Hear from a customer that uses AWS very differently and has the ability to move rapidly to address pressing societal problems with the aid of AWS. Learn how mobile and IoT technology allows them to scale and accelerate their impact. Speakers talk about why they selected specific AWS services, and how that allowed their teams to be agile and focus more on their solutions instead of undifferentiated plumbing. The ShotSpotter Flex gunfire data service uses acoustic sensors to detect gunshots in the field. Data is then transmitted to the AWS Cloud, and further processed to validate signatures and, if necessary, alert emergency dispatch teams.
Georgie Mathews - Senior Product Manager, Amazon Web Services & Raj Behara - Sr Director, Agero
Amazon Pinpoint, a new AWS service, makes it easy to run targeted campaigns to improve user engagement. Pinpoint helps you understand app user behavior, define who to target, what push notification to send, when to deliver the notifications, and track results.
Rahul Sharma - GM, AWS Mobile, Amazon Web Services & Sudeep Kulkarni - Cloud Software Engineer, New York Life Labs
In this session, we dive deep into how to apply the serverless microservices approach to developing mobile and web applications on AWS, for both business-to-consumer and business-to-employee enterprise applications. We discuss use cases, scenarios, best practices, and design patterns around user authentication, authorization, backend microservices, API management, analytics, user engagement, security, and integration with enterprise apps and data sources. We go into details about how to use AWS Lambda, Amazon API Gateway, Amazon Cognito, and other AWS services to develop mobile applications and backend microservices.
Justin Pirtle - Senior Solutions Architect, Amazon Web Services & Jim Tran - Principal Enterprise SA, Amazon Web Services
By leveraging 'serverless architectures', startups and enterprises are building and running modern applications and services with increased agility and simplified scalability—all without managing a single server. Many applications need to manage user identities and support sign-in/sign-up. In this session, we dive deep on how to support millions of user identities, as well as how to integrate with social identity providers (such as Google and Facebook) and existing corporate directories. You learn the real-world design patterns that AWS customers use to implement authentication and authorization. By combining Amazon Cognito identity pools and user pools with API Gateway, AWS Lambda, and AWS IAM, you can add security without adding servers.
Rohan Deshpande - Sr. Software Engineer, Amazon Web Services
Amazon Echo and Alexa have shown that voice interfaces provide significant benefits to users – interactions are easy, fast, and context-driven. In this hands-on session, you’ll see how to add compelling voice and chat interfaces to your mobile apps, using Amazon Lex for processing conversations and triggering corresponding actions in your backend systems, all without having to manage any infrastructure. You’ll leave knowing how to build apps that can “Find me a nearby hotel” or “Reorder supplies for the copier”.
Vikram Madan - Senior Technical Product Manager, Amazon Web Services & Tim Hunt - Sr. Product manager, Amazon Web Services & Ravi Tiyyagura - Sr Director, Enterprise Architecture, Asurion
Secure user sign-up and sign-in is critical for many mobile and web applications. Amazon Cognito is the easiest way to secure your mobile and web applications by providing a comprehensive identity solution for end user management, registration, sign-in, and security. In this product deep dive, we will walk through Cognito’s feature set, which includes serverless flows for user management and sign-in, a fully managed user directory, integrations with existing corporate directories, and many other features. In addition, we will cover key use cases and discuss the associated benefits.
Ilya Epshteyn - Solutions Architect, Amazon Web Services & Sudeep Kulkarni - Cloud Software Engineer, New York Life Labs & Rohit Katti - CTO - NYL Labs, New York Life Insurance Co
Insurance processes are fragmented and manual in nature. For the next generation of digitally savvy consumers, New York Life Labs uses modern technology and analytics to engage with and provide best-in-class customer service. This session will discuss how New York Life Labs is mobilizing key insurance technology components including customer service, underwriting, policy administration, and claims processing. This session will focus on how New York Life Labs is taking a mobile-first approach in redesigning claims processing with responsive design, automated workflows, and AWS mobile services. Some of the services discussed will be Amazon Cognito, Amazon SNS, AWS Lambda, Amazon DynamoDB, and Amazon SWF, and how they all come together.
Andrew Chud - Senior Software Development Engineer, Amazon Web Services
This session takes a practical approach to developing real-world enterprise applications (business-to-consumers and business-to-employees) using a serverless backend that can scale to virtually unlimited users without any infrastructure to manage. Learn how to develop enterprise apps using AWS Mobile Hub, Amazon Cognito, Amazon API Gateway, and AWS Lambda to implement best practices for authentication/authorization, cloud logic, and secure integration of existing enterprise resources and user directories. Finally, see how to employ a robust application lifecycle (build, test, and deliver) and implement a multi-stage rollout to production.
Richard Threlkeld - Solutions Architect - Mobile, Amazon Web Services
In this deep-dive session, we outline how to leverage the appropriate AWS services for sending different types and sizes of data, such as images or streaming video. We'll cover common real-world scenarios related to authentication/authorization, access patterns, data transfer and caching for more performant Mobile Apps. You learn when you should access services such as Amazon Cognito, Amazon DynamoDB, Amazon S3, or Amazon Kinesis directly from your mobile app, and when you should route through Amazon API Gateway and AWS Lambda instead. Additionally, we cover coding techniques across the native, hybrid, and mobile web using popular open-source frameworks to perform these actions efficiently, and with a smooth user experience.
Becky Weiss - Senior Principal Engineer, Amazon Web Services
In this session, we walk through the fundamentals of Amazon VPC. First, we cover build-out and design fundamentals for VPC, including picking your IP space, subnetting, routing, security, NAT, and much more. We then transition into different approaches and use cases for optionally connecting your VPC to your physical data center with VPN or AWS Direct Connect. This mid-level architecture discussion is aimed at architects, network administrators, and technology decision-makers interested in understanding the building blocks that AWS makes available with Amazon VPC and how you can connect this with your offices and current data center footprint.
Sean Meckley - Sr. Product Manager, Amazon Web Services & Vahram Sukyas - VP, Application Infrastructure & Operations, Warner Bros.
Whether you’re running a simple website, a mobile app, or a suite of business applications, DNS is a fundamental part of any architecture in the cloud. In this mid-level architecture session, we’ll cover everything you need to get started with Amazon Route 53, AWS’s highly-available DNS service. You’ll learn how to use public DNS, including routing techniques such as weighted round-robin, latency-based routing, and geo DNS; how to configure DNS failover using health checks; how and when to use private DNS within your Virtual Private Cloud (VPC); and how Amazon Route 53 interacts with Amazon EC2’s DNS for instance naming and DNS resolution across your network.
Peter Dalbhanjan - Solutions Architect, Amazon Web Services & Kit Ewbank - Lead Cloud Engineer, Capital One & Jeff Storey - Engineering Tools Lead, Capital One
Capital One began moving to AWS just two years ago. Every day, the amount of traffic we serve from the cloud continues to grow. With development teams having the freedom to choose their own technology stacks, many teams have quickly started moving applications to Docker. In this session, learn how Capital One uses a combination of the Elastic Load Balancing service along with Application Load Balancer features to increase deployment speed and reliability.
Alan Halachmi - Sr. Manager, Solutions Architecture, Amazon Web Services
Recently, AWS announced support for Internet Protocol version 6 (IPv6) for several AWS services, providing significant capabilities for applications and systems that need IPv6. This session provides an overview of IPv6 and covers key aspects of AWS support for the protocol. We discuss Amazon S3 and S3 Transfer Acceleration, Amazon CloudFront and AWS WAF, Amazon Route 53, AWS IoT, Elastic Load Balancing, and the virtual private cloud (VPC) environment of Amazon EC2. The presentation assumes solid knowledge of IPv4 and those AWS services.
Dan Bowling - Director, Global Cloud Ecosystem, Level 3 Communications, L.L.C. & Michael Dews - Dir, Product Management, Level 3 Communications, L.L.C.
You can leverage the agility and scale of cloud services and consolidate your data centers, but performance and security are only as good as your WAN. Is this session, learn best practices for connecting cloud, data center, and branch sites through public and private networking to maximize performance, minimize costs, and simplify security.
Amy Machi - Director, AT&T & Chris Porell - Enterprise Solution Consultant, AT&T Global Business Solutions
Learn how the AT&T MPLS VPN with the network of tomorrow’s virtualized network functions and Software Defined Networking (SDN) will help you create and deliver agile workloads for your Enterprise. You’ll also learn how AT&T combines trending viability of open standards-based software for broader network applications. Additionally, you’ll see how the AT&T NetBond API integration with AWS Direct Connect removes the complexity and enables on-demand, private connection within minutes via a self-service portal. AT&T NetBond connects your people, your data, and your business directly to your AWS services. This fast, highly secure, scalable, private network connection increases performance, while improving control and delivering a better ROI for your enterprise applications. Join us for an informative session on how you can enhance your cloud connectivity with AT&T and AWS.
sergey royt - Software Development Manager, Amazon Web Services
As companies grow and expand their global footprint, it becomes increasingly critical to make systems highly available while also improving responsiveness to end-users. Companies are choosing to place their applications closer to end-users to improve performance, which introduces the complications of how to route end-user traffic to the most appropriate endpoints and how to most efficiently route traffic within internal systems.
Kaartik Viswanath - Product Manager, Amazon Web Services & Camil Samaha - Senior Manager Solutions Architecture, Amazon Web Services
Amazon’s Virtual Private Cloud (Amazon VPC) lets you provision a logically isolated section of the Amazon Web Services (AWS) cloud and gives you complete control over your virtual networking environment. Amazon VPC continues to evolve with new capabilities and enhancements. These features give you increasingly greater isolation, control, and visibility at the all-important networking layer. In this session, we review some of the latest changes, discuss their value, and describe their use cases.
Andrew Braham - Manager - Cloud Network Engineering, Netflix & Laurie Ferioli - Senior Program Manager, Netflix
Netflix was one of the earliest very large AWS customers. By 2014, we were running hundreds of applications in Amazon EC2. That was great, until we needed to move to VPC. Given our scale, uptime requirements, and the decentralized nature of how we manage our production environment, the VPC migration (still ongoing) presented particular challenges for us and for AWS as it sought to support our move. In this talk, we discuss the starting state, our requirements and the operating principles we developed for how we wanted to drive the migration, some of the issues we ran into, and how the tight partnership with AWS helped us migrate from an EC2-Classic platform to an EC2-VPC platform.
sidhartha chauhan - Solutions Architect, Amazon Web Services & Benjamin Feldon - Enterprise Solutions Architect, Amazon Web Services
Many enterprises on their journey into the cloud require consistent and highly secure connectivity between their existing data center and AWS footprints. In this session, we walk through the different architecture options for establishing this connectivity using AWS Direct Connect and VPN. With each option, we evaluate the considerations and discuss risk, performance, encryption, and cost. As we walk through these options, we try to answer some of the most common questions that typically arise from enterprises that tackle design and implementation. You'll learn how to make connectivity decisions that are suitable for your workloads, and how to best prepare against business impact in the event of failure.
Mingxue Zhao - Sr Product Manager - Technical, Amazon Web Services & Mark Ryland - Director, Solutions Architecture, Amazon Web Services
In this session, we explore AWS support for IPv6 for full end-to-end connectivity for EC2 instances inside of a VPC. IPv6 on EC2 instances introduces new capabilities and interesting new wrinkles into the VPC model. Customer VPCs receive IPv6 addresses from an Amazon address block, and existing features such as security groups, network ACLs, route tables, peering, and gateways have been enhanced to support IPv6. Finally, we look at some future capabilities planned for IPv6 in VPC.
David Pippenger - Sr Manager, TechOps, GREE International, Inc.
Gree is a gaming company with the distinction of launching the world’s first mobile social game in 2007 and later went on to develop a rich portfolio of original games. With an active acquisition and development strategy, Gree is transforming the way games are built. Learn how Gree leveraged Amazon VPCs and Aviatrix Enterprise cloud-defined networking in AWS Marketplace to rapidly integrate acquisitions and develop new games to maintain competitive differentiation in this fast moving segment. Hear about the use case, requirements and lessons learned. Learn how you can transitioned to Amazon VPCs in the middle of the development process.
Steve Mueller - Specialist Solutions Architect, Amazon Web Services
In this session, we walk through the Amazon VPC network presentation and describe the problems we were trying to solve when we created it. Next, we walk through how these problems are traditionally solved, and why those solutions are not scalable, inexpensive, or secure enough for AWS. Finally, we provide an overview of the solution that we've implemented and discuss some of the unique mechanisms that we use to ensure customer isolation, get packets into and out of the network, and support new features like VPC endpoints.
Steve Seymour - Solutions Architect, Amazon Web Services
As enterprises move to the cloud, robust connectivity is often an early consideration. AWS Direct Connect provides a more consistent network experience for accessing your AWS resources, typically with greater bandwidth and reduced network costs. This session dives deep into the features of AWS Direct Connect and VPNs. We discuss deployment architectures and demonstrate the process from start to finish. We show you how to configure public and private virtual interfaces, configure routers, use VPN backup, and provide secure communication between sites by using the AWS VPN CloudHub.
David Brown - Director, Software Development, Amazon Web Services
Elastic Load Balancing automatically distributes incoming application traffic across multiple Amazon EC2 instances for fault tolerance and load distribution. In this session, we go into detail about Elastic Load Balancing configuration and day-to-day management, as well as its use in conjunction with Auto Scaling. We explain how to make decisions about the service and share best practices and useful tips for success.
Mike Furr - Principal Engineer, Amazon Web Services
Many applications are network I/O bound, including common database-based applications and service-based architectures. But operating systems and applications are often untuned to deliver high performance. This session uncovers hidden issues that lead to low network performance, and shows you how to overcome them to obtain the best network performance possible.
Colm MacCarthaigh - Principal Engineer, Amazon Web Services
Launched in June of 2015, s2n is an AWS open-source implementation of the TLS and SSL network security protocols, which focus on security, simplicity, and performance. With development led by engineers from Amazon EC2, Amazon S3, Amazon CloudFront, and AWS security and cryptography services, s2n is a unique opportunity to observe how we develop and test security and availability for critical software at AWS. Learn how we iterate and code, how we automate software verification beyond the usual code reviews, and how open source works at Amazon.
SAC: Security & Compliance
Mark Nunnikhoven - Vice President, Cloud Research, Trend Micro & Robert Hotaling - CISO, Cengage Learning & Karla Krause-Miller - Manager, Marketing Events
Does meeting stringent compliance requirements keep you up at night? Do you worry about having the right audit trails in place as proof?
Chris Colthurst - Solutions Architect, Amazon Web Services & Heitor Vital - Solutions Architect, Amazon Web Services & Vlad Vlasceanu - Sr. Solutions Architect, Amazon Web Services
In this workshop, we help you understand how you can help protect your web applications from threats cost effectively by using AWS WAF and Amazon CloudFront. As attacks and attempts to exploit vulnerabilities in web applications become more sophisticated and automated, having an effective web request filtering solution becomes key to keeping your users' data safe. We will cover common attack vectors and what you can do to mitigate them. You will learn how to leverage AWS WAF in conjunction with Amazon CloudFront to detect unwanted traffic and block it using simple configurations and automations.
Jeff Wierer - Sr. Manager, Amazon Web Services
Are you interested in learning how to control access to your AWS resources? Have you ever wondered how to best scope down permissions to achieve least privilege permissions access control? If your answer to these questions is 'yes', this session is for you. We take an in-depth look at the AWS Identity and Access Management (IAM) policy language. We start with the basics of the policy language and how to create and attach policies to IAM users, groups, and roles. As we dive deeper, we explore policy variables, conditions, and other tools to help you author least privilege policies. Throughout the session, we cover some common use cases, such as granting a user secure access to an Amazon S3 bucket or to launch an Amazon EC2 instance of a specific type.
Michael Capicotto - Solutions Architect, Amazon Web Services & Matt Nowina - Solutions Architect, Amazon Web Services
In a rapidly changing IT environment, detecting and responding to new threats is more important than ever. This session shows you how to build a predictive analytics stack on AWS, which harnesses the power of Amazon Machine Learning in conjunction with Amazon Elasticsearch Service, AWS CloudTrail, and VPC Flow Logs to perform tasks such as anomaly detection and log analysis. We also demonstrate how you can use AWS Lambda to act on this information in an automated fashion, such as performing updates to AWS WAF and security groups, leading to an improved security posture and alleviating operational burden on your security teams.
Chad Woolf - Director, AWS Risk & Compliance, Amazon Web Services & Sara Duffer - Director Security Assurance Automation, Amazon Web Services
Is your IT environment getting bigger and more complex than your compliance team can handle? Get a peek under the hood of how the AWS Compliance team manages and automates security assurance and compliance in the AWS environment. We’ll tell you what we’re doing to automate controls, match up huge data sets to validate compliance, how we perform game day simulations of entire region outages, and how we manage our ever-present external audits. With each example, we’ll give you some ideas on how to use AWS services to manage the security and compliance of your AWS and on-prem environments. In this session, Chad Woolf, Director of Risk and Compliance for AWS, and Sara Duffer, Director of Security Assurance Automation discusses how the AWS Compliance team uses AWS services like Amazon Inspector, Amazon CloudWatch Logs, AWS CloudTrail, and AWS Config to manage risk, compliance, and audit in the massive scale of the AWS IT environment.
Eric Brandwine - Sr. Principal Engineer, Amazon Web Services
Encryption is a favorite of security and compliance professionals everywhere. Many compliance frameworks actually
Jason Chan - Engineering Director, Netflix
Historically, relationships between developers and security teams have been challenging. Security teams sometimes see developers as careless and ignorant of risk, while developers might see security teams as dogmatic barriers to productivity. Can technologies and approaches such as the cloud, APIs, and automation lead to happier developers and more secure systems? Netflix has had success pursuing this approach, by leaning into the fundamental cloud concept of self-service, the Netflix cultural value of transparency in decision making, and the engineering efficiency principle of facilitating a “paved road.”
Aaron Klein - Co-Founder/COO, CloudCheckr & Aaron Newman - Co-Founder/CEO, CloudCheckr
CloudCheckr Co-Founders Aaron Newman and Aaron Klein will highlight effective strategies and tools that AWS users can employ to improve their security posture. Specific emphasis will be placed upon leveraging native AWS services and the talk will include concrete steps that users can begin employing immediately. Session sponsored by CloudCheckr.
Scott Pack - Sr Security Engineer, Adobe & Randy Young - Principal Product Manager, Splunk
Ensuring security and compliance across a globally distributed, large-scale AWS deployment requires a scalable process and a comprehensive set of technologies. This session will deep-dive into the AWS native monitoring and security services and some Splunk technologies leveraged globally to perform security monitoring across a large number of AWS accounts. You will learn about the collection plumbing including components of S3, Kinesis, CloudWatch, SNS, Dynamo DB and Lambda, as well as the tooling and processes used at Adobe to deliver scalable monitoring without managing an unwieldy number of API keys and input stanzas. Session sponsored by Splunk.
Dave Walker - Specialist Solutions Architect, Security and Compliance, Amazon Web Services & James Brown - VP, Technical Solutions Group, Alert Logic
AWS serverless architecture components such as Amazon S3, Amazon SQS, Amazon SNS, CloudWatch Logs, DynamoDB, Amazon Kinesis, and Lambda can be tightly constrained in their operation. However, it may still be possible to use some of them to propagate payloads that could be used to exploit vulnerabilities in some consuming endpoints or user-generated code. This session explores techniques for enhancing the security of these services, from assessing and tightening permissions in IAM to integrating tools and mechanisms for inline and out-of-band payload analysis that are more typically applied to traditional server-based architectures.
Chris McCurdy - Solutions Architect Specialist in Healthcare and Lifesciences, Amazon Web Services & Alan Nihill - IT Analyst, Johnson & Johnson
Johnson & Johnson is in the process of doing a proof of concept to rewrite the compliance framework that they presented at re:Invent 2014. This framework leverages the newest AWS services and abandons the need for continual describes and master rules servers. Instead, Johnson & Johnson plans to use a distributed, event-based architecture that not only reduces costs but also assigns costs to the appropriate projects rather than central IT.
Hart Rossman - Global Practice Manager, Security, AWS Professional Services, Amazon Web Services & Bill Shinn - Principal Security Solutions Architect, Amazon Web Services
This session tells how our most mature, security-minded Fortune 500 customers adopt AWS while improving end-to-end protection of their sensitive data. Learn about the enterprise security architecture decisions made during actual sensitive workload deployments as told by the AWS professional services and the solution architecture team members who lived them. In this very prescriptive, technical walkthrough, we share lessons learned from the development of enterprise security strategy, security use-case development, security configuration decisions, and the creation of AWS security operations playbooks to support customer architectures.
Balaji Palanisamy - Senior Consultant - Security, Amazon Web Services & Rolando Glogowiec - Global Director, Information Security, Time Inc.
Professional services has completed five deep PCI engagements with enterprise customers over the last year. Common patterns were identified and codified in various artifacts. This session introduces the patterns that help customers address PCI requirements in a standard manner that also meets AWS best practices. Hear customers speak about their side of the journey and the solutions that they used to deploy a PCI compliance workload.
Chris Whalley - Security Assurance TPM, Amazon Web Services & Sujatha Guruswamy - Director, Cloud & Web Services, Merck & Daniel Dziadiw - Director, IT Compliance and Risk Management, Merck
GxP is an acronym that refers to the regulations and guidelines applicable to life sciences organizations that make food and medical products such as drugs, medical devices, and medical software applications. The overall intent of GxP requirements is to ensure that food and medical products are safe for consumers and to ensure the integrity of data used to make product-related safety decisions. The term GxP encompasses a broad range of compliance-related activities such as Good Laboratory Practices (GLP), Good Clinical Practices (GCP), Good Manufacturing Practices (GMP), and others, each of which has product-specific requirements that life sciences organizations must implement based on the 1) type of products they make and 2) country in which their products are sold. When life sciences organizations use computerized systems to perform certain GxP activities, they must ensure that the computerized GxP system is developed, validated, and operated appropriately for the intended use of the system. For this session, co-presented with Merck, services such as Amazon EC2, Amazon CloudWatch Logs, AWS CloudTrail, AWS CodeCommit, Amazon Simple Storage Service (S3), and AWS CodePipeline will be discussed with an emphasis on implementing GxP-compliant systems in the AWS Cloud.
Prashant Prahlad - Principal Product Manager, Amazon Web Services & Stefan Lenz - Senior Key Expert Datadriven Services, Cloud and Virtualisation, Siemens AG
This session enables security operators to automate governance and implement use cases addressed by AWS services such as AWS CloudTrail, AWS Config Rules, Amazon CloudWatch Events, and Trusted Advisor. Based on the nature of vulnerabilities, internal processes, compliance regimes, and other priorities, this session discusses the service to use when. We also show how to detect, report, and fix vulnerabilities, or gain more information about attackers. We dive deep into new features and capabilities of relevant services and use an example from an AWS customer, Siemens AG, about how to best automate governance and scale. A prerequisite for this session is knowledge of security and basic software development using Java, Python, or Node.
Venkat Vijayaraghavan - Sr. Product Manager, Amazon Web Services
As attackers become more sophisticated, web application developers need to constantly update their security configurations. Static firewall rules are no longer good enough. Developers need a way to deploy automated security that can learn from the application behavior and identify bad traffic patterns to detect bad bots or bad actors on the Internet. This session showcases some of the real-world customer use cases that use machine learning and AWS WAF (a web application firewall) with automated incident response and machine learning to automatically identify bad actors. We also present tutorials and code samples that show how customers can analyze traffic patterns and deploy new AWS WAF rules on the fly.
Zaher Dannawi - Sr. Product Manager -- Technical, Amazon Web Services
This session covers AWS Identity and Access Management (IAM) best practices that can help improve your security posture. We cover how to manage users and their security credentials. We also explain why you should delete your root access keys—or at the very least, rotate them regularly. Using common use cases, we demonstrate when to choose between using IAM users and IAM roles. Finally, we explore how to set permissions to grant least privilege access control in one or more of your AWS accounts.
Martin Sirull - DevOps Consultant, Amazon Web Services & Mirza Baig - Sr Director, Software Engineering, Experian
This session covers what a real-world production deployment of a fully automated deployment pipeline looks like with instances that are deployed without SSH keys. By leveraging AWS CloudFormation along with Docker and AWS CodeDeploy, we show how we achieved semi-immutable and fully immutable infrastructures, and what the challenges and remediations were.
Sam Elmalak - Solutions Architect, Amazon Web Services & Michael Capicotto - Solutions Architect, Amazon Web Services & Adam Fokken - Principal Architect, ThomsonReuters
Whether it is per business unit or per application, many AWS customers use multiple accounts to meet their infrastructure isolation and billing requirements. In this session, we discuss considerations, limitations, and security patterns when building out a multi-account strategy. We explore topics such as identity federation, cross-account roles, consolidated logging, and account governance. Thomson Reuters also presents their approach to a multi-account strategy.
Josh Du Lac - Senior Consultant, Amazon Web Services & Matt Bretan - Senior Security Consultant, Amazon Web Services
Now that you have decided to move forward with an AWS multi-account strategy, how should you help secure and manage your accounts? This session is a deep dive into some of the tools and processes that you can use to help ensure that you have a hardened blast radius between your AWS accounts while still being able to manage your security baseline across them. Some of the topics we cover are:
Misha Govshteyn & Sven Skoog - Information Security Officer, Monotype Imaging
In this session, you’ll learn how to setup your AWS environment to avoid a cyber security attack and how to build a cloud specific incident response plan if your organization is breached. Cyber security expert and founder of Alert Logic Misha Govshteyn will share lessons learned from organizations that have protected themselves in AWS and from those that have been breached but quickly resolved their issues and implemented strong controls as a result. Joining Misha will be Sven Skoog, Senior Manager IT Security at Monotype who will discuss the cyber security posture they implemented within the AWS cloud and how they have built a robust process to ensure ongoing protection. Session sponsored by Alert Logic.
Andrew Thomas - Director, Anti-DDoS and WAF, Amazon Web Services & Prasad Kalyanaraman - VP Edge Services, Amazon Web Services
At re:Invent 2016, we are launching AWS Shield, a managed DDoS protection service. With AWS Shield, you can help protect Amazon CloudFront, Elastic Load Balancing, and Amazon Route 53 resources from DDoS attacks. In addition to introducing AWS Shield, this session presents some of the things we do behind the scenes to detect and mitigate Layer 3/4 network attacks and highlights ways you can use this new service to protect against Layer 7 application attacks.
Anders Samuelsson - Principal TPM, Amazon Web Services
AWS Organizations is a new administrative capability, which allows you to control multiple AWS accounts centrally. With Organizations, you can hierarchically organize and manage your AWS accounts and apply organizational controls across these accounts to meet your business needs. In this session, we cover the capabilities of AWS Organizations and discuss best practices when managing multiple AWS accounts.
Leo Zhadanovsky - Principal Solutions Architect, Amazon Web Services & Thomas Vachon - Manager of Cloud Architecture, Harvard University
With constantly evolving threats across the Internet, Harvard University deployed a security network platform to mitigate cyber threats, current and future, to protect institutional and research data. By using multiple geographic locations, best of breed equipment, and network automation, Harvard provides visibility, availability, and multilayer protections for their cloud network. This talk discusses the benefits, considerations, and lessons learned from using their security network platform at the edge of the cloud. Learn how Harvard designed and deployed the platform, utilizing serverless architecture to orchestrate the solution from within to protect their most sensitive data and afford students, faculty, and staff the flexibility of cloud computing.
Mike Wagers - Justic and Public Safety Lead, Amazon Web Services & Ben Potter - Security Consultant, Amazon Web Services & John Fokker - Digital Team Coordinator, NHTCU & Raj Samani
Come to this session to learn how Europol, the Dutch police, Intel Security, and Kaspersky Lab have come together in an unprecedented collaboration of government and private-sector organizations. Open source ransomware code makes it easier to lock victims’ computers and encrypt their data, resulting in an alarming increase of cyber ransom. In response www.nomoreransom.org was created with the additional cooperation of AWS and Barracuda Web Application Firewall. Learn what tools are available to retrieve encrypted data and take a peek under the hood of this mission-critical website in the fight against ransomware. Perhaps because the site opposes ransomware, it has already received a number of attacks. Learn how AWS and Intel worked to rebuff these persistent assaults.
Henrik Johansson - Security Solutions Architect, Amazon Web Services
This session demonstrates 5 different security and compliance validation actions that you can perform using Amazon CloudWatch Events and AWS Config rules. This session focuses on the actual code for the various controls, actions, and remediation features, and how to use various AWS services and features to build them. The demos in this session include CIS Amazon Web Services Foundations validation; host-based AWS Config rules validation using AWS Lambda, SSH, and VPC-E; automatic creation and assigning of MFA tokens when new users are created; and automatic instance isolation based on SSH logons or VPC Flow Logs deny logs. This session focuses on code and live demos.
John Martinez - Head of Products and Customer Solutions, Evident.io & Tim Prendergast & Anthony Johnson - Staff Engineer, Ellie Mae
We are all embarking on a journey in the cloud that can be frightening at times, thrilling at others, but at all times filled with pitfalls and scary monsters that threaten the security of our infrastructure, applications, and data. The ultimate reward for all our hard work is to achieve a state of autonomous, self-healing security within our environment--one that can withstand any threats, whether internal or external. In this session, we walk you through the steps you need to be successful in your journey, just like Ellie Mae and many other enterprises and agencies. Your journey starts with security automation, and from there you will push outside of your security comfort zone, thanks to the gift of enhanced visibility and omniscience. Next we use CloudFormation Templates and custom signatures to move through our next security challenge with speed, and finally, we build auto-remediation into our security strategy with AWS Lambda workflows that enable the system to self-correct when misconfigurations occur. This fast-paced session will be filled code, best practices to help you in your quest, and even a few surprises about the ultimate destination of your journey.
SEC: Security Services Mini-Conference
Blake Frantz - Sr. Technical Program Manager, Amazon Web Services & Rob Witoff - Director, Coinbase
Audit teams can consistently evaluate the security of an AWS account. Best practices greatly reduce complexity when managing risk and auditing the use of AWS for critical, audited, and regulated systems. You can integrate these security checks into your security and audit ecosystem. Center for Internet Security (CIS) benchmarks are incorporated into products developed by 20 security vendors, are referenced by PCI 3.1 and FedRAMP, and are included in the National Vulnerability Database (NVD) National Checklist Program (NCP). This session shows you how to implement foundational security measures in your AWS account. The prescribed best practices help make implementation of core AWS security measures more straightforward for security teams and AWS account owners.
Becky Weiss - Senior Principal Engineer, Amazon Web Services
AWS IAM and Amazon VPC offer powerful tools that help you adhere to the principle of least privilege in your resource permissions and network security settings. This workshop will start with the fundamentals of IAM and VPC security techniques and will give you hands-on experience in writing, testing, applying, troubleshooting, and auditing progressively more tightly scoped IAM policies. You will also get experience building and monitoring VPC security groups that grant only the access required to perform tasks.
Ken Beer - General Manager, AWS Key Management Service, Amazon Web Services & Cory Minkovich - Staff Software Engineer, Box Inc
AWS Key Management Service provides an easy and cost-effective way to secure your data in AWS. In this session, you learn about leveraging the latest features of the service to minimize risk for your data. We also review the recently released Import Key feature that gives you more control over the encryption process by letting you bring your own keys to AWS.
Bryan Miller - Solutions Builder, Amazon Web Services
This session shows you how to reduce your blast radius by using multiple AWS accounts per region and service, which helps limit the impact of a critical event such as a security breach. Using multiple accounts helps you define boundaries and provides blast-radius isolation. Though managing multiple accounts can be difficult, we will present an upcoming AWS solution that will help automate the process for controlling cross- account access by managing roles across multiple accounts.
Eugene Yu - Senior Consultant, Amazon Web Services & Eric Gifford - Information Security Architect, Cambia Health Solutions & Brad Davidson - Security Engineer, Cambia Health Solutions
Cloud computing offers many advantages, such as the ability to scale your web applications or website on demand. But how do you scale your security and compliance infrastructure along with the business? Join this session to understand best practices for scaling your security resources as you grow from zero to millions of users. Specifically, you learn the following:
Quint Van Deman - Sr. Consultant, IT Transformation, Amazon Web Services
AWS supports identity federation using SAML (Security Assertion Markup Language) 2.0. Using SAML, you can configure your AWS accounts to integrate with your identity provider (IdP). Your federated users then are authenticated and authorized by your organization's IdP, and they can use single sign-on (SSO) to access AWS.
Andrew Flavell - VP - Architecture, Nike
Microservice architectures provide numerous benefits but also have significant security challenges. This session presents how Nike uses layers of security to protect consumers and business. We show how network topology, network security primitives, identity and access management, traffic routing, secure network traffic, secrets management, and host-level security (antivirus, intrusion prevention system, intrusion detection system, file integrity monitoring) all combine to create a multilayer, end-to-end security solution for our microservice-based premium consumer experiences. Technologies to be covered include Amazon Virtual Private Cloud, access control lists, security groups, IAM roles and profiles, AWS KMS, NAT gateways, ELB load balancers, and Cerberus (our cloud-native secrets management solution).
Moataz Anany - Solutions Architect, Amazon Web Services & Nate Sammons - Principal Architect, Nasdaq
Security of big data workloads in a hybrid IT environment often comes as an afterthought. This session discusses how enterprises can architect secure, big-data workloads on AWS. We cover the application of authentication, authorization, encryption, and additional security principles and mechanisms to workloads leveraging Amazon Elastic MapReduce and Amazon Redshift.
Alexander Lucas - Principal Security Engineer, Amazon Web Services
Attend this session to learn about security testing your applications in AWS. Effective security testing is challenging, but multiple features and services within AWS make security testing easier. This session covers common approaches to testing, including how we think about testing within AWS, how to apply AWS services to your test setup, remediating findings, and automation.
Jeff Lyon - Technical Operations Manager, Amazon Web Services & Andrew Kiggins - Security Solutions Architect, Amazon Web Services & Adrian Newby - CTO, Crownpeak & David Grampa - Owner, TypeFrag.com
Distributed denial of service (DDoS) attack mitigation has traditionally been a challenge for those hosting on fixed infrastructure. In the cloud, users can build applications on elastic infrastructure that is capable of mitigating and absorbing DDoS attacks. What once required overprovisioning, additional infrastructure, or third-party services is now an inherent capability of many cloud-based applications. This session explains common DDoS attack vectors and how AWS customers with different use cases are addressing these challenges. As part of the session, we show you how to build applications that are resilient to DDoS and demonstrate how they work in practice.
Brigid Johnson - Senior Product Manager - Technical, Amazon Web Services
Managing permissions across a growing number of identities and resources can be time-consuming and complex. Testing, validating, and understanding permissions before and after policy changes are deployed is critical to ensuring that your users and systems have the appropriate level of access. This session walks through the tools that are available to test, validate, and understand the permissions in your account. We demonstrate how to use these tools and how to automate them to continually validate the permissions in your accounts. The tools demonstrated in this session help you answer common questions such as:
Stephen Schmidt - VP and Chief Information Security Officer, Amazon Web Services & Marene Allison - Chief Information Security Officer, Johnson & Johnson
AWS CISO Steve Schmidt presents the state of the union for re:Source Mini Con for Security Services. He addresses the state of the security and compliance ecosystem; large enterprise customer additions in key industries; the vertical view: maturing spaces for AWS security assurance (GxP, IoT, CIS foundations); and the international view: data privacy protections and data sovereignty. The state of the union also addresses a number of new identity, directory, and access services, and closes by looking at what's on the horizon.
Don Bailey - Principal Security Engineer, Amazon Web Services & Brian Wagner - Security Consultant, Amazon Web Services
With security-relevant services such as AWS Config, VPC Flow Logs, Amazon CloudWatch Events, and AWS Lambda, you now have the ability to programmatically wrangle security events that may occur within your AWS environment, including prevention, detection, response, and remediation. This session covers the process of automating security event response with various AWS building blocks, taking several ideas from drawing board to code, and gaining confidence in your coverage by proactively testing security monitoring and response effectiveness before anyone else does.
Chris Whalley - Security Assurance TPM, Amazon Web Services & Mitsuhiro YANO - Senior Planner, Information Solution, Sysmex Corporation
This session is intended for customers seeking to build out a comprehensive plan around data integrity in the cloud. Information governance, system validation, PHI and PII records come with their own set of regulatory considerations - and by using tools such as AWS CloudFormation, Amazon Virtual Private Cloud, Directory Service, AWS SDKs, RDS, AWS CloudTrail, and Amazon CloudWatch portions of the governance burden can be lifted, offloaded, or partnered around. A look at the top ten data integrity controls will be considered, including a customer presentation featuring an example of a regulated quality management system. Finally, the physical import/export tool Snowball's addition to the AWS Business Associate Agreement (BAA) program will be discussed, along with the practical application and security implications of chain of custody.
Byron Cook - Senior Principal Security, Amazon Web Services
Automatic and semiautomatic mechanical theorem provers are now being used within AWS to find proofs in mathematical logic that establish desired properties of key AWS components. In this session, we outline these efforts and discuss how mechanical theorem provers are used to replay found proofs of desired properties when software artifacts or networks are modified, thus helping provide security throughout the lifetime of the AWS system. We consider these use cases:
Mai-Lan Tomsen Bukovec - Vice President and General Manager, Amazon Web Services
This session provides a foundational overview of the AWS storage portfolio, including block, file, object, and cloud data migration services. This session will touch on the significant new offerings, outline some of the most common use cases and prepare you for the individual deep dive sessions, customer sessions and new announcements.
Edward Naim - Head of Product, Amazon EFS, Amazon Web Services & Darryl Osborne - Solutions Architect - Storage, Amazon Web Services
In this session, we fill you in about Amazon EFS, including an overview of this recently introduced service, its use cases, and best practices for working with it.
Jed Sundwall - Global Open Data Lead, Amazon Web Services & Shay Har-Noy - VP & General Manager, Platform, DigitalGlobe & John Webb - Manager, Information Technology, E & J Gallo Winery
Making earth observation data available by using Amazon S3 is accelerating scientific discovery and enabling the creation of new products. Attend and learn how the scale and performance of Amazon S3 lets earth scientists, researchers, startups, and GIS professionals gather and analyze planetary-scale data without worrying about limitations of bandwidth, storage, memory, or processing power. Learn how AWS is being used to combine satellite imagery, social data, and telemetry data to produce new products and services. Learn also how Amazon S3 provides much more than storage, and how an open geospatial data lake on Amazon S3 can be used as the basis for planetary-scale applications built with Amazon EMR, Amazon API Gateway, and AWS Lambda. As part of this talk, AWS customer Digital Globe demonstrates how they use open data stored in S3 to distribute high-resolution satellite imagery to their customers around the world.
Brad Steele - Global Manager, Venture Capital BD, Amazon Web Services & Armando Wirshing - Senior Business Development Manager, Forty8Fifty Labs / Veristor & Will White - CIO, Mapbox & Jason Gluckman - Lead Software Engineer, Ring & Rob Hruska - Engineering Director, Hudl
Join us for this lightning-round showcase of hot new brands and startup companies that are using AWS to play a really big game. You'll hear from experts like Mapbox CIO Will White, Ring Senior Engineer Jason Gluckman, Hudl Engineering Director Rob Hruska, and many others as they explain how they thought about the problems they faced and how they solved them in this TED-style session packed with lots of creative thinking.
Sajee Mathew - Principal Solutions Architect, Amazon Web Services & Austin Fonacier - Lead Software Architect, Spokeo
At Spokeo, we are running a fast, big data, and high-traffic website providing people search services. But at our scale, we started to reach limitations to how fast our conventional web stack could do things and concluded that a Ruby on Rails–only solution simply couldn't keep up. In this session, we cover some of the options we had to solve this problem and why we chose Amazon Elastic File System (Amazon EFS) as a central part of our solution with metrics and benchmarking. Using EFS, we were able to take response times down from 250 ms to below 70 ms. We look into the architecture of the solution and lessons we learned along the way. In the end, we find that faster response times are just the beginning of the benefits that we see.
Darryl Osborne - Solutions Architect - Storage, Amazon Web Services & Brad Bressler - Technical Account Manager, Atlassian & Neal Riley - Principle Solutions Engineer, Atlassian
At Atlassian, we create popular software tools to help every team unleash their full potential. We use our issue tracking tool JIRA to handle customer support issues from around the globe. As the business grew, we decided to move from a single server instance of JIRA to our JIRA Data Center product on AWS infrastructure to increase reliability, scale, and security. JIRA Data Center's requirement to use a shared file system caused us to try several in-house and third-party solutions prior to the availability of Amazon Elastic File System (Amazon EFS). We chose EFS for its performance, ease of use, automatic scaling, and out-of-the-box distribution capabilities. In this session, we discuss how Atlassian uses EFS to run JIRA Data Center. Topics include the technical architecture, how and why we chose EFS, our recommendations, and the lessons learned along the way.
Alexander Crosett III - Enterprise Solutions Architect, Amazon Web Services & Carrington Wong - Platform Solutions Engineer, Monsanto & Vishnu Alavur Kannan - Analytics Technical Platforms Lead, Monsanto
At Monsanto, we build and use technologies that support our data and also BI efforts that facilitate intelligent, data-driven decisions. In the past year, we've embarked on large-scale efforts to modernize our geospatial platform and improve our analytic processing capabilities by building out new cloud and open-source based services. We found using Amazon Elastic File System (Amazon EFS) gave us the flexibility and performance we were seeking while saving us significant time, effort, and cost. In this session, we discuss how Monsanto uses the Amazon EFS service to run our large scaling geospatial data sets such as raster, and to perform highly parallelized analytics for data scientists and business users. Topics include the technical architecture, how and why we chose EFS for handling data sets that are terabytes in size, our recommendations, and the lessons learned along the way.
Henry Zhang - Senior Product Manager, Amazon Web Services & Rich Sutton - VP of Engineering, Digital Risk, Social, Proofpoint
Without careful planning, data management can quickly turn complex with a runaway cost structure. Enterprise customers are turning to the cloud to solve long-term data archive needs such as reliability, compliance, and agility while optimizing the overall cost. Come to this session and hear how AWS customers are using Amazon Glacier to simplify their archiving strategy. Learn how customers architect their cloud archiving applications and share integration to streamline their organization's data management and establish successful IT best practices.
Marc Trimuschat - Head of Storage Technical Business Development, Amazon Web Services & Hari Tatrakal - Director Of Database Services, Live Nation Entertainment
As the global leader of live entertainment, Live Nation promotes and produces over 22,000 events annually, operates out of 37 countries, and cultivates over 530 million fans globally. To focus on the growth of the business and shed increasing infrastructure costs, the company made the strategic decision to get out of the data center business and go all in with the cloud. Using instrumental services like AWS Import/Export Snowball, VM Import/Export, AWS CloudFormation and AWS Identity and Access Management, VP Cloud Services Jake Burns quickly and efficiently migrated priority business and operational applications, allowing for immediate cost efficiencies. Learn how AWS offerings like Snowball played a decisive role in Live Nation's ability to easily migrate data and enable end users to quickly access applications to minimize operational impact.
Mike Kuentz - Senior Solutions Architect & Benjamin Snively - Solutions Architect, Amazon Web Services & Steve Elliott - AWS Enterprise Education Acct Exec, Amazon Web Services & Casey Helfrich - Chief Architect, UPMC & Stuart Ingram - Principle Architect, UPMC
Today's health care systems generate massive amounts of protected health information (PHI) — patient electronic health records, imaging, prescriptions, genomic profiles, insurance records, even data from wearable devices. In this session, UPMCe dives deep into two efforts: Their 'Data Liberation Project' — a next-gen petabyte-scale software solution that provides responsible management of PHI within their own environments as well as externally, and “Neutrino” a real time medical document aggregator which utilizes natural language processing techniques to unlock hidden value from unstructured narratives. UPMC Enterprises (UPMCe), a division of University of Pittsburgh Medical Center, builds technology and invests in health care companies, from new startups to large established partners, with an eye toward revolutionizing healthcare. They embody the startup mentality with a focus on innovation and creating new data-heavy applications—all in support of new spin-off companies, furthering economic development, and disrupting healthcare. Join us to learn how they do security management and governance using Amazon S3, Amazon EC2, AWS Config, AWS CloudTrail, and other Amazon services help UPMCe think big about healthcare data in the public sector.
Sabina Joseph - Head of Storage Global Partnership and Alliances, Amazon Web Services & Steven Nelson - Global Storage Segment Technical Lead, Amazon Web Services & Jeff Pisano - VP Global operations, Carlyle Group & Mark Buckley - Senior Engineer, General Dynamics IT & Richard Spurlock - CEO and Founder, Cobalt Iron
Customers seeking to make transformations to their environment typically gain efficiencies in both operational and capital costs. This session presents three different case studies, each from a different category of operations – Private Equity, Managed Services, and Government. In Government, General Dynamics will discuss how implementing AWS storage as part of their DevOps environment, saving $700k per year. Carlyle Group will show how replacing their on-premises storage with cloud based solutions brought new capabilities to their organization. Finally, Cobalt Iron will discuss how moving their customers to cloud backup targets allowed them to deliver 2M cloud backups per month.
Asa Kalavade - GM AWS Storage Gateway, Amazon Web Services & Paul Reed - Snr. Product Manager, Amazon Web Services & Ferrol Macon - VP Architecture and Product Strategy, VeriStor
AWS Storage Gateway seamlessly combines existing infrastructure investment with the AWS Cloud for backup, restore, disaster recovery, hybrid workloads, and tiered storage use cases. It has a new file interface that supports moving files in NFS-based file systems into Amazon S3. This session dives into the use cases and advantages of hybrid cloud storage, exploring this newly updated data transfer service with stories from real-life customer deployments.
Colin Lazier - General Manager, Glacier & Snowball, Amazon Web Services & Rich Ridolfo - Sr. Director, Operations & Infrastructure R & D, Philips & Jay Littlepage - VP, Infrastructure & Operations, DigitalGlobe
At re:Invent 2016 AWS announced several revolutionary ways to move data into and out of the AWS Cloud in large batches, so you can retire servers, racks, and data centers and embark on new projects faster. Two of these new services are AWS Snowmobile and AWS Snowball Edge. Snowmobile cuts cloud data migration projects from years to weeks in time with a 45 foot truck with 100PB of capacity. AWS Snowball Edge is a petabyte-scale data transfer service with portable storage and compute capabilities. It transports up to 100TB of data in a ruggedized, weather resistant, rack mountable, shippable device with the same embedded cryptography and security as the original Snowball. Learn more about each of these services and hear from two customers on the impact these services have made on their data center operations.
Omair Gillani - Sr. Product Manager, Amazon Web Services & John Elliott - Capacity Planning Engineer, Pinterest
Customers using Amazon S3 at large scale benefit greatly from storage management features. Storage lifecycle policies help them reduce storage costs. Cross-region replication makes it easier to copy data between AWS regions for compliance or disaster recovery. Event notifications allow automatic initiation of processes on objects as they arrive, or capture information about objects and log it for security purposes. In this session, you'll learn about these features, and also several new storage management features in Amazon S3 that give users unmatched visibility into what data they are storing and how that data is being used. These new features make it simpler to analyze usage by users, apps, or organizations, to highlight anomalies, and to optimize business process workflows. They also help identify opportunities to reduce costs, improve performance, and archive infrequently used data. In addition, they can provide insight into who is accessing data stored in S3. As part of this talk, AWS customer Pinterest shows how they have been able to leverage many of the new S3 storage management features to reduce their storage costs significantly by moving a large amount of their data from S3 Standard to S3 Standard – Infrequent Access storage.
Uri Wolloch - CEO, N2W Software & Tommy Johnston - Sr. Cloud Administrator, Aptean
Aptean is a global enterprise software provider that uses AWS as the core of its infrastructure because it’s a solution that reliably backs up Aptean’s Amazon EC2 instances. Come to this session to learn what happened when Aptean needed a highly reliable, full backup solution that also allowed for ease of scale, automation, and instant recovery in case of a failure. In this session, you’ll learn how CPM helps take full advantage of AWS Snapshots, thus adding a management layer to control retention, automate recoveries, and allow live, application-aware backup of both Windows and Linux instances. Aptean will also why they selected N2W Cloud Protection Manager (CPM) in AWS Marketplace as its backup solution extending AWS services and supporting their diverse customer base. By the end of the session, you’ll have learned the details of how CPM helps Aptean properly manage its snapshots and recoveries, ensuring a resilient deployment that meets Aptean’s business continuity goals.
Rob Alexander - AWS Solutions Architect, Amazon Web Services
In this popular session, you will learn about the latest features and use cases for Amazon EBS, including best practices, an overview of newly introduced features, and brand-new re:Invent announcements. In particular we will cover the expanded portoflio of volume types, including provisioned IOPS, cold storage, and throughput-optimized. This session will help database admins and application architects understand how to blend performance and cost with applicaitns for big data analytics, data warehousing, and transactional and NoSQL databases.
Masataka Kubo - Senior Product Manager, Amazon Web Services & Andy Shenkler - EVP and Chief Solutions & Technology Officer, Sony DADC New Media Solutions
Not just for archiving or compliance use cases, Amazon Glacier accommodates customers simply looking to replace their on-premises long term storage with a cost efficient, durable, cloud option, from which they can easily and quickly access their data when they need to. This session will introduce newly launched features for Amazon Glacier, review the current service feature set, and share the global data center shut down and storage strategy for Sony DADC New Media Solutions (NMS). NMS is Sony’s digital servicing division providing global digital distribution, linear playout and white label OTT/Commerce solutions for clients such as BBC Worldwide, NBCUniversal, Sony Playstation, and Funimation Entertainment.
Susan Chan - Senior Product Manager, Amazon S3, Amazon Web Services
Come learn about new and existing Amazon S3 features that can help you better protect your data, save on cost, and improve usability, security, and performance. We will cover a wide variety of Amazon S3 features and go into depth on several newer features with configuration and code snippets, so you can apply the learnings on your object storage workloads.
Shaown Nandi - VP, Head of Infrastructure and Cloud, Dow Jones & News Corp & John Fox - Principal Technical Product Manager, Commvault & Peter Laudenslager - Director of Product Marketing, Commvault
These days, EVERY workload is considered critical by someone in the organization. As a result, SLAs are shrinking. IT is challenged to meet these SLAs, but there isn’t enough budget to provide services like disaster recovery (DR) using traditional methods and infrastructure. The good news is that public cloud platforms, like AWS, are becoming the de facto infrastructure choice for DR. However, workload portability solutions that simplify cross-platform or cloud recovery are required to meet most RTO & RPO SLAs in the cloud. AWS provides the infrastructure we need to bring DR to tier 2 and tier 3 workloads that have never been able to afford it before. Now, we need orchestration and automation to make it scalable and reliable.
Srikanth Devidi - Sr. Data Engineer, Netflix Inc. & Albert Wong - Enterprise Reporting Platform Manager, Netflix
You have billions of events in your fact table, all of it waiting to be visualized. Enter Tableau… but wait: how can you ensure scalability and speed with your data in Amazon S3, Spark, Amazon Redshift, or Presto? In this talk, you’ll hear how Albert Wong and Srikanth Devidi at Netflix use Tableau on top of their big data stack. Albert and Srikanth also show how you can get the most out of a massive dataset using Tableau, and help guide you through the problems you may encounter along the way.
Zoltan Nagy - Infrastructure Tech-Stack Owner, Prezi & Tamas Nemeth - Data Platform Tech Lead, Prezi
Prezi has over 75 million registered users generating 900 GB of new data each day. They store XML describing over 260 million public presentations that have been viewed over 2 billion times. This created a scale problem that on-premises storage couldn't solve. In this session, you'll learn how Prezi leveraged the power and flexibility of Amazon S3 to turn their storage problem into an analytics opportunity and scale storage to meet the demands of their business without overspending. Prezi provides insight on how a small team of engineers accepted the challenge and succeeded using AWS with managed Hadoop (Amazon EMR) and Amazon Redshift. You'll see how Prezi runs and improves its data platform in a self-service data culture with this small team, without drowning in maintenance and support. Additionally, Prezi shows how the infrastructure team architected their services for storing and serving crucial data, user-generated and otherwise, using S3 and Amazon CloudFront—including pitfalls, best practices, cost considerations, and learnings along the way.
John Hitchingham - Director of Performance Engineering, FINRA
FINRA partnered with AWS product teams to leverage Amazon EMR and Amazon S3 extensively to build an advanced analytics solution. In this session, you'll hear how FINRA implemented a data lake on S3 to provide a single source for their big data analytics platform. FINRA ingests 75 billion records each day of stock market transactions, with an AWS storage footprint of 20 petabytes across S3 and Amazon Glacier. To deal with this workload, FINRA has architected a platform that separates storage from compute to manage capacity for each independently, leading to improved performance and cost effectiveness. You'll also learn how FINRA was able to leverage Hbase on Amazon EMR to achieve significant benefits over running Hbase on a fixed capacity cluster. FINRA was able to implement a system that seamlessly scales in response to data growth and can scale quickly in response to user traffic. By working with multiple clusters, FINRA can now isolate ETL and user query workloads and has achieved rapid, built-in disaster recovery capability by leveraging data storage on S3 to run from multiple AZs and across regions.
Kevin Stinson - Senior Engineer, Quantcast & David Hanson - Director of Infrastructure, Smartsheet.com
Startups around the world use AWS services to access the power of the cloud to grow faster and more cost effectively. In this session, Smartsheet talks about how they were able to cost-effectively build their prototype for scale and avoid replatforming at different points in the adoption curve, and Quantcast discusses how they are running a high-performance analytics solution on AWS. They provide several tips and tricks for S3, and show how they removed a traditional MySQL data store from a distributed-image hosting application so that the only required data store is S3. They also show how to avoid common, cumbersome database practices by working with the eventually consistent nature of S3 objects and the fact that objects and directories share the same namespace.
David Stein - Business Development Manager, Amazon Web Services & Kyle House - Dev Ops Engineer, ZenDesk & Dave Bernstein - Sr. Director, Cloud Operations, Zendesk, ZenDesk & Paul Frederiksen - Principal DevOps Engineer, Videology & David Ortiz - Senior Software Engineer, Videology
The companies Videology and Zendesk both had the same problem—how to rearchitect their big data processing platforms to scale to meet growing demand, while at the same time improving performance, availability, and cost structure? Videology provides a converged advertising solution that is screen-agnostic, ensuring unduplicated reach with the right frequency cadence to achieve guaranteed results. To achieve this, their big data platform ingests, processes, and analyzes a variety of logs. In this talk, Videology discusses how they migrated our platform to use Cloudera on a mix of m4 and r3 instances using the Amazon EBS Streaming Optimized HDD (st1) volume type. Zendesk provides a cloud-based customer support platform that allows quicker and easier interaction between businesses and customers. To deliver this experience, Zendesk runs a large Elasticsearch, Logstash, Kibana stack. The talk also discusses how Zendesk rearchitected their deployment to use m4s and also leverage the EBS Streaming Optimized HDD (st1) volume type. Tips for success will be shared throughout.
SVR: Serverless Computing Mini-Conference
Tim Bray - Senior Principal Engineer, Amazon Web Services
AWS Step Functions is a new, fully-managed service that makes it easy to coordinate the components of distributed applications and microservices using visual workflows. Step Functions is a reliable way to connect and step through a series of AWS Lambda functions so that you can build and run multi-step applications in a matter of minutes. This session shows how to use AWS Step Functions to create, run, and debug cloud state machines to execute parallel, sequential, and branching steps of your application, with automatic catch and retry conditions. We share how customers are using AWS Step Functions to reliably scale multi-step applications such as order processing, report generation, and data transformation–all without managing any infrastructure.
Tim Wagner - General Manager, Amazon Web Services
In this session, you'll learn what’s new and hot with AWS Lambda. Come on a tour with Dr. Tim Wagner, General Manager of AWS Lambda, to learn what we’ve been working on and what we are planning for the future. You'll get a hands-on demonstration of some our newest features which will provide you with a launching pad for some of the later sessions in the day.
Cecilia Deng - Software Developer, Amazon Web Services & Marco Pierleoni - Lead Software Engineer, Thomson Reuters & Anders Fritz - Senior Manager, Strategic Product Management, Thomson Reuters
Serverless architecture can eliminate the need to provision and manage servers required to process files or streaming data in real time. In this session, we will cover the fundamentals of using AWS Lambda to process data in real-time from push sources such as AWS Iot and pull sources such as Amazon DynamoDB Streams or Amazon Kinesis. We will walk through sample use cases and demonstrate how to set up some of these real-time data processing solutions. We'll also discuss best practices and do a deep dive into AWS Lambda real-time stream processing.
Balaji Iyer - AWS Professional Service Consultant, Amazon Web Services & Rao Prahlad - Solutions Architect, Amazon Web Services
AWS Lambda empowers developers to build cloud-native web applications or platforms using microservices architecture. The three-tier web app has been a cornerstone architecture pattern for decades in this space. This session walks you through the process of identifying the presentation, logic, and data tiers required to build web applications with AWS Lambda at the core. By using AWS Lambda as your logic tier, you have a wide number of data storage options for your data tier. AWS offers a wide range of database services to fit your application requirements. We dive into methodologies for picking the right database technology based on your application requirements.
Andrew Baird - Solutions Architect, Amazon Web Services & Michael Connor - Director of Technical Strategy, The Coca-Cola Company & Patrick Brandt - Solutions Architect, The Coca-Cola Company
The value of building serverless applications on AWS applies to businesses of all sizes. For enterprises that own hundreds or thousands of applications, the cost of managing and securing servers could equal millions of dollars per year. In this session, we will discuss how building serverless applications on AWS can eliminate many of those costs and still meet enterprise requirements across a breadth of areas: including identity and access control, network security, performance, availability, deployment and operations. Also, Coca-Cola North America will give a glimpse into the tools and strategies they’ve developed in order to make their next generation of digital marketing applications completely serverless, and give a peak into the innovative ways they’re planning to use AWS Lambda in the future.
Vishal Singh - Senior Product Manager, Amazon Web Services
Chatbots are the new apps. Businesses of all sizes, from startups to enterprises, are looking for new ways to connect with their users through natural, conversational interfaces. Developers have started using chatbots to improve the productivity and efficiency of their operations. In this session, we show how to use AWS Lambda and other serverless offerings from AWS to build chatbots quickly and efficiently. We share examples from our recently concluded AWS Slack Hackathon with a full walkthrough of building a conversational chatbot in an easy, fast, and fun way, along with helpful tools, tips, and techniques.
Vyom Nagrani - Manager, Product Management, Amazon Web Services & Richard McFarland - VP Data Services and Chief Data Scientist, Hearst Corp
Ever wished you had a list of cheat codes to unleash the full power of AWS Lambda for your production workload? Come learn how to build a robust, scalable, and highly available serverless application using AWS Lambda. In this session, we discuss hacks and tricks for maximizing your AWS Lambda performance, such as leveraging customer reuse, using the 500 MB scratch space and local cache, creating custom metrics for managing operations, aligning upstream and downstream services to scale along with Lambda, and many other workarounds and optimizations across your entire function lifecycle.
Kuldeep Chowhan - Principal Engineer, Expedia Inc
In the middle of 2015, Expedia started using AWS Lambda for serverless computing. We built boilerplate templates in Node.js, Java, and Python so development teams could build and deploy serverless applications into AWS. Currently, we have 300 AWS Lambda functions processing 40 million invocations per day.
Orr Weinstein - Senior Product Manager, AWS Lambda, Amazon Web Services
Amazon API Gateway and AWS Lambda provide a new way of building applications by removing servers from the picture. But what does the removal of servers mean to tasks like deployment, monitoring, and debugging? How should you set up blue-green deployments or set alarms? Come learn all this and more, including how to use AWS services and tools like AWS CodePipeline, AWS CloudFormation, and Amazon CloudWatch to manage your serverless applications at high quality.
Miguel Alvarado - VP of Data and Analytics, Vevo & Alan Zawari - Senior Engineer, Content Services, Vevo
Vevo has undergone a complete strategic and technical reboot, driven not only by product, but also by engineering. Since November 2015, Vevo has been replacing monolithic, legacy content services with a modern, modular, microservices architecture, all while developing new features and functionality. In parallel, Vevo has built its data platform from scratch to power internal analytics as well as a unique music video consumption experience through a new personalized feed of recommendations — all in less than one year.
Brittany Doncaster - Solutions Architect, Amazon Web Services & Andrew Baird - Solutions Architect, Amazon Web Services
Wild Rydes (www.wildrydes.com) needs your help! With fresh funding from its seed investors, Wild Rydes is seeking to build the world’s greatest mobile/VR/AR unicorn transportation system. The scrappy startup needs a first-class webpage to begin marketing to new users and to begin its plans for global domination. Join us to help Wild Rydes build a website using a serverless architecture. You’ll build a scalable website using services like AWS Lambda, Amazon API Gateway, Amazon DynamoDB, and Amazon S3. Join this workshop to hop on the rocket ship!
Olivier Klein - Emerging Technologies Solutions Architect, Asia-Pacific, Amazon Web Services
Bots are eating the world! Wild Rydes (
Tim Wagner - General Manager, Amazon Web Services & Tim Griesbach - Sr. Director Technology, FINRA
Join us to learn about the state of serverless computing from Dr. Tim Wagner, General Manager of AWS Lambda. Dr. Wagner discusses the latest developments from AWS Lambda and the serverless computing ecosystem. He talks about how serverless computing is becoming a core component in how companies build and run their applications and services, and he also discusses how serverless computing will continue to evolve.
Randall Hunt - Developer Evangelist, Amazon Web Services
Defining infrastructure resource policies in an organized manner can help your company better manage its infrastructure resources.
Bob Kinney - Senior SDE, Amazon Web Services
In this session, you learn how to monitor and manage your serverless APIs in production. We show you how to set up Amazon CloudWatch alarms, interpret CloudWatch logs for Amazon API Gateway and AWS Lambda, and automate common maintenance and management tasks on your service.
Zlatan Dzinic - Senior Architect, Amazon Web Services
Learn how to architect fully available and scalable Microsoft solutions and environments in AWS. Find out how Microsoft solutions can leverage various AWS services to achieve more resiliency, replace unnecessary complexity, simplify architecture, provide scalability, introduce DevOps concepts, automation and repeatability. Plan authentication and authorization, various hybrid scenarios with other cloud environment and on premise solutions/infrastructure. Learn about common architecture patterns for Active Directory and business productivity solutions like SharePoint, Exchange and Skype for Business, also common scenarios for SQL deployments and System Center.
OLGA LAGUNOVA - Chief Data and Analytics Officer, Pitney Bowes & Kyle Pause - Director of SaaS Platform Development, Pitney Bowes
Pitney Bowes is reinventing its business based on a SaaS and a cloud-based model to deliver services for their clients globally centered on the Pitney Bowes Commerce Cloud. The
Sal Chiovari - CIO, Edwards Lifesciences & Rajeev Bhardwaj - Sr. Director, Enterprise Architecture, Edwards Lifesciences
Edwards Lifesciences partners with clinicians to develop innovative technologies in the areas of structural heart disease and critical care monitoring to help patients live longer, healthier and more productive lives. The Edwards Lifesciences IT team recently embarked on a journey to move many of their business critical corporate IT applications to AWS to better support the Edwards Lifesciences mission to help patients worldwide. In this session, the Edwards Lifesciences team would like to propose a different approach for Corporate IT teams to consider with respect to the cloud that can help fulfill key business outcomes while transforming IT infrastructure quickly… and safely! An approach that starts with business critical applications to take full advantage of the scale and growth that were key requirements for the team. It worked for us and created a new vibe in our IT organization which we are currently enjoying.
Adnan Ijaz - Principal Product Manager, Amazon Web Services
Today, we are announcing EC2 Systems Manager. Amazon EC2 Systems Manager is a management service that helps you automatically collect software inventory, apply OS patches, create system images, and configure Windows and Linux operating systems. These capabilities help you define and track system configurations, prevent drift, and maintain software compliance of your EC2 and on-premises configurations. This session provides an overview of these newly announced services and how they work together within the larger AWS ecosystem to provide comprehensive management capabilities.
Scott Zimmerman - Solutions Architect, Amazon Web Services
Running Microsoft workloads on AWS is easy and can save you money. This session will cover how to bring your own Microsoft licenses to AWS, and then demonstrate using PowerShell to import your Windows Server image from Vmware or Hyper-V, configure Windows KMS with your license key, and launch an EC2 Dedicated Host. We will discuss ways you can use AWS Config rules to manage license compliance.
WIN303: How to Launch a 100K-User Corporate Back Office with Microsoft Servers and AWS
Bill Jacobi - Senior Solutions Architect, Amazon Web Services
Learn how to build a scalable, compliance-ready, and automated deployment of the Microsoft “backoffice” servers for 100K users running on AWS. In this session, we show a reference architecture deployment of Exchange, SharePoint, Skype for Business, SQL Server and Active Directory in a single VPC. We discuss the following: (1) how the solution is automated for 100K users, (2) how the solution is enabled for compliance (e.g., FedRAMP, HIPAA, PCI), and (3) how the solution is built from modular 10K user blocks. Attendees should have knowledge of AWS CloudFormation, PowerShell, instance bootstrapping, VPCs, and Amazon Route 53, as well as the relevant Microsoft technologies.
Zlatan Dzinic - Senior Architect, Amazon Web Services & Lou De La Torre - Solutions Architect, Amazon Web Services
AWS can help you rapidly deploy and scale your Microsoft SharePoint environment to help you collaborate more efficiently and cost-effectively. This session reviews architectural considerations for building a SharePoint deployment on AWS, best practices to ensure optimal performance, how to leverage multiple Availability Zones for high availability and disaster recovery, and how to integrate with Microsoft Active Directory. We will also look at new Quick Start guides, AWS CloudFormation templates, and other tools that dramatically reduce the time to deployment.
Ron Cully - Sr. Product Manager, Amazon Web Services
Active Directory (AD) is essential for Windows workloads in the cloud. AWS offers customers multiple ways to integrate AD with cloud workloads like EC2, RDS, and AWS Enterprise Applications: AWS Directory Service for Microsoft Active Directory (Enterprise Edition) as a managed service and Active Directory running on AWS EC2 Windows instances. Which option is right for you? This session will discuss the key deployment considerations for each option to help you identify which best meets your project goals, and the effort involved. The session will cover options for integrating with your on-premises directory, port and security considerations, application considerations, and best practices.
Vlad Vlasceanu - Sr. Solutions Architect, Amazon Web Services
This session goes deep on best practices and considerations for running Microsoft SQL Server on AWS. We cover best practices for deploying SQL Server, how to choose between Amazon EC2 and Amazon RDS, and ways to optimize the performance of your SQL Server deployment for different application types. We review in detail how to provision and monitor your SQL Server databases, and how to manage scalability, performance, availability, security, and backup and recovery in both Amazon RDS and Amazon EC2. In addition, we cover how you can set up a disaster recovery solution between an on-premises SQL Server environment and AWS, using native SQL Server features like log shipping, replication, and AlwaysOn Availability Groups.
Ari Bixhorn - VP Technical Evangelism, Panopto & Timothy Sullivan - VP Engineering, Panopto
Video is a 'last-mile problem' for search technology. Unlike webpages, documents, and email, content in videos has traditionally been impossible to search. Recent advances in automated speech and text recognition, however, let businesses and universities search inside video assets as easily as inside textual content. In this session, you'll learn how Panopto is using AWS to solve the video-search problem at scale, while saving over 50% in operating costs by taking advantage of Spot instances. We discuss the cross-platform architecture that combines Windows and Linux to provide cost-effective video processing and search indexing. We also dive deep into scaling Spot elastically based on user demand, handling fallback situations when instances are revoked, and using the Spot bidding process to optimize cost structure. Finally, we discuss future plans to reduce operating costs even further through Spot fleets and grid processing.
Taylor Anderson - Sr. Product Manager, Amazon Web Services & Amjad Hussain - Senior Manager, Amazon Web Services
Whether you are a traditional enterprise making a shift to the cloud or are already all-in, AWS EC2 management capabilities enable you to perform common tasks of inventory collection, and patch and image management of your infrastructure at scale. In this session, we'll provide a brief overview of these powerful capabilities and bring it all together with a demo of real-world hybrid-cloud management scenario.
Amjad Hussain - Senior Manager, Amazon Web Services & Tim Nicholas - Product Owner and Lead Automation Architect, Xero Ltd
In this session, you will learn how you can use EC2 management capabilities to perform repeatable automation of your infrastructure at scale, across platforms and hybrid environments. As you embark on your cloud journey and embrace a modern DevOps mindset, you not only want to deploy software quickly but also ensure configuration consistency at cloud scale. Many enterprises have successfully used services such as EC2 Run Command to perform administrative tasks, and we'll share some of those stories. In addition, we will demo new capabilities to ensure a desired state approach to software configuration either through predefined or easy to build custom configurations, and also how you can integrate with other AWS services to deliver enterprise IT and business value.
Scott McDonald - Cloud Infrastructure Architect, Amazon Web Services & Zlatan Dzinic - Senior Architect, Amazon Web Services
Migrating on-premises workloads to AWS is a top priority for customers today. AWS has an array of tools and capabilities to make the move faster and less risky. This deep-dive session presents how to build a migration framework for your Windows-based applications by using AWS CloudFormation templates, along with multisite replication capabilities built into AWS and optimized for Microsoft Windows workloads built on SQL HA. This framework helps speed up your data center migration and minimize risk during the process.
WWPS: World Wide Public Sector
Christopher Gile - Sr Mgr, Security Assurance, AWS Security, Amazon Web Services & Duane Shugars - COO, EITC, Inc.
Learn how to architect for compliance in the AWS cloud and see how your organization can leverage the agility, cost savings, scalability, and flexibility of the cloud while meeting the most stringent regulatory and compliance requirements, including Federal Risk and Authorization Management Program (FedRAMP), ITAR, CJIS, HIPAA, and DoD Cloud Computing Security Requirements Guide (SRG) Levels 2 and 4. Hear best practices and practical use cases for using AWS GovCloud (US) to comply with a variety of regulatory regimes.
Blake Chism - Practice Manager, Amazon Web Services & Dawn Beadle - Director, Infrastructure Services, Monash University & Alison Robinson - Associate Vice President & Deputy CIO, University of Maryland
Cloud governance is critical to executing a systemic, supportable, and sustainable cloud transformation strategy. This session explores best practices to achieve that model, including standards policies, automation that consistently applies and enforces policies and controls, self-service capabilities that enable agility and speed, and automated monitoring and cost management that ensure operational integrity. A well-developed cloud governance model enables customers to improve operational integrity, reliability, performance, and transparency using the AWS cloud. This session guides you through a best practice governance model, including policy considerations and recommendations, self-service automation methods towards IT-as-a-Service, and use case examples.
Doug VanDyke - Director, Amazon Web Services & Eric Hysen - Executive Director DHS Digital Service, DHS Digital Service & Marina Martin - Chief Technology Officer, Department of Veterans Affairs (VA) & Lt Gen Samuel Greaves - Commander, Space and Missile Systems Center (SMC), U.S. Air Force & Chris Lynch - Director, Defense Digital Service (DDS) & Eric Schoonover - Digital Service Expert, Defense Digital Service (DDS)
Cloud computing can help government organizations increase innovation, efficiency, agility, and resiliency—all while reducing costs. This session highlights perspectives on how small, highly empowered United States Digital Service (USDS) teams within the Department of Defense (DoD), Department of Homeland Security (DHS), and Department of Veterans Affairs (VA) are breaking down innovation barriers, tackling mission-critical operations, and delivering more value with the cloud in highly regulated, unclassified environments.
Brad Dispensa - Sr. Solutions Architect, Amazon Web Services & Elizabeth Boudreau - Director of Information Technology, Claritas Genomics & Gerard Gallant - CJIS Security Specialist, Motorola Solutions & Frank Shuang Chen - Software Architect, Coursera
With rich controls, auditing, and broad security accreditations, AWS enables its customers to be in compliance with CJIS, FERPA, and HIPAA. Come hear customers share their approaches to achieving compliance for these and other standards across many markets.
Harish Doraiswamy - Research Scientist, New York University & Fernando Chirigati - Doctoral Candidate and Research Assistant, NYU Tandon School of Engineering
In this session, learn how Data Polygamy, a scalable topology-based framework, can enable users to query for statistically significant relationships between spatio-temporal datasets. With the increasing ability to collect data from urban environments and a push toward openness by governments, we can analyze numerous spatio-temporal datasets covering diverse aspects of a city. Urban data captures the behavior of the city’s citizens, existing infrastructure (physical and policies), and environment over space and time. Discovering relationships between these datasets can produce new insights by enabling domain experts to not only test but also generate hypotheses. However, discovery is difficult. A relationship between two datasets can occur only at locations or time periods that behave differently compared to the regions’ neighborhood. The size and number of datasets and diverse spatial and temporal scales at which the data is available presents computational challenges. Finally, of several thousand possible relationships, only a small fraction is actually informative. We have implemented the framework on Amazon EMR and show through an experimental evaluation using over 300 spatial-temporal urban datasets how our approach is scalable and effective at identifying significant relationships. Find details about the work at
Teresa Carlson - Vice President, Worldwide Public Sector, Amazon Web Services & Alison Robinson - Associate Vice President & Deputy CIO, University of Maryland & Devika Chawla - Director of Engineering, Netflix & Tony Prophet - Chief Equality Officer, Salesforce & Leah Gilliam - VP of Education and Innovation, Girls Who Code
In this session, learn about efforts to diversify IT, from programs in high schools all the way up to diversity among top-level executives. Regardless of where you are in your career, this session dives deep to speak to you about the changing tech landscape. Through successes and failures, speakers candidly discuss how everyone can help in changing the ratio to attract a more diverse workforce and create a culture of inclusion. Bring your questions, network with peers, and be inspired by both women and men making an impact in non-profit, enterprise and government sectors.