Computing Systems are undergoing a multitude of interesting changes: from the platforms (cloud, appliances) to the workloads, data types, and operations (big data, machine learning). Many of these changes are driven or being tackled through innovation in hardware even to the point of having fully specialized designs for particular applications. In this talk I will review some of the most important changes happening in hardware and discuss how they affect system design as well as the opportunities they create. I will focus on data processing with an emphasis on streams and event based systems but also discuss applications in other areas. I will also briefly discuss how these trends are likely to result in a very different form of IT, and consequently of Computer Science, from the one we know today.
Gustavo Alonso is a Professor of Computer Science at ETH Zürich. He studied telecommunications -electrical engineering- at the Madrid Technical University (ETSIT, Politécnica de Madrid). As a Fulbright scholar, he completed an M.S. and Ph.D. in Computer Science at UC Santa Barbara. After graduating from Santa Barbara, he worked at the IBM Almaden Research Center before joining ETHZ. His research interests encompass almost all aspects of systems, from design to run time. He works on distributed systems, data processing, and system aspects of programming languages. Most of his research these days is related to multi-core architectures, data centers, FPGAs, and hardware acceleration. Gustavo has received numerous awards for his research, including three Test-of-Time awards for work in databases, programming languages, and systems. He is a Fellow of the ACM and of the IEEE as well as a Distinguished Alumnus of the Department of Computer Science of UC Santa Barbara.
Quantitative models in informatics are well accepted and employed, whereas qualitative, conceptual, modeling is frequently considered with scepticism: huge effort is said to trade for poor benefit.
In this talk I discuss criteria for adequate conceptual models of event driven systems, and discuss what kind of benefit such models may yield.
In particular, I will argue that an adequate model supports not only understanding, abstraction, structure, behavior, and user interaction. In addition, a lot of nontrivial, deep insight can be gained, including invariants in a broad sense, with e.g. models in physics as a role model.
Wolfgang Reisig is a professor emeritus at the Informatics Institute of Humboldt-Universitaet zu Berlin, Germany.
After his PhD at RWTH Aachen in 1979, he joined the group of Carl Adam Petri at Gesellschaft fuer Mathematik und Datenverarbeitung (GMD), and served as a professor at Technical University of Munich, 1987-1993, from where he moved to Berlin.
Prof. Reisig was a senior research at the International Computer Science Institute (ICSI) in Berkeley, California in 1997, got the "Lady Davis Visiting Professorship" at the Technion, Haifa (Israel), the Beta Chair of Technical University of Eindhoven, and twice received an IBM Faculty Award for his contribution to Cross-organizational Business Processes and the Analysis of Service Models. He has been the speaker of a PhD school on Service Oriented Architectures, 2010 - 2017.
Prof. Reisig is a member of a member of the European Academy of Sciences, Academia Europaea. He published and edited numerous books and articles on Petri Net Theory and Their Applications, as well as other modeling techniques.
In the last four years, stream processing has gone from niche to mainstream, with real-time data processing systems gaining traction in not only fast-moving startups, but also their more skeptical and cautious enterprise brethren. In light of such pervasive adoption, is it safe to say we’ve we finally reached the point where stream processing is a solved commodity? Are we done and ready to move on to the next big thing?
In this talk, I will argue that the answer to those questions is conclusively "no": stream processing as a field of research is alive and well. In fact, as streaming systems evolve to look more and more alike, the need for active exploration of new ideas is all the more pressing. And though streaming systems are more capable and robust than ever, they remain in many ways difficult to use, difficult to maintain, and difficult to understand. But we can change that.
I don’t claim to have all the answers; no one does. But I do have a few ideas of where we can start. And by sharing my thoughts on some of the more interesting open problems in stream processing, and encouraging others to share theirs, I’m hoping that we as a research community can work together to help move the needle just a little bit further.
Tyler Akidau is a software engineer at Google, where he is the technical lead for the Data Processing Languages & Systems group, responsible for Google's Apache Beam efforts, Google Cloud Dataflow, and internal data processing tools like Google Flume, MapReduce, and MillWheel. He is also a founding member of the Apache Beam PMC, author of the 2015 Dataflow Model paper, the Streaming 101 and Streaming 102 articles, and co-author of the Streaming Systems book. Though deeply passionate and vocal about the capabilities and importance of stream processing, he is also a firm believer in batch and streaming as two sides of the same coin, with the real endgame for data processing systems the seamless merging between the two.
Online event processing are abound ranging from web and mobile applications to data processing. Such event processing applications often require the ability to ingest, store, dispatch and process events. Until now, supporting all of these needs has required different systems for each task -- stream processing engines, messaging queuing middleware, and pub/sub messaging systems. This has led to the unnecessary complexity for the development of such applications and operations leading to increased barrier to adoption in the enterprises.
In this keynote, Karthik will outline the need to unify these capabilities in a single system and make it easy to develop and operate at scale. Karthik will delve into how Apache Pulsar was designed to address this need with an elegant architecture. Apache Pulsar is a next generation distributed pub-sub system that was originally developed and deployed at Yahoo and running in production in more than 100+ companies. Karthik will explain how the architecture and design of Pulsar provides the flexibility to support developers and applications needing any combination of queuing, messaging, streaming and lightweight compute for events. Furthermore, he will provide real life use cases how Apache Pulsar is used for event processing ranging from data processing tasks to web processing applications.
Karthik Ramasamy is the co-founder and CEO of Streamlio that focuses on building next generation event processing infrastructure using Apache Pulsar. Before Streamlio, he was the engineering manager and technical lead for real-time infrastructure at Twitter where he co-created Twitter Heron. Twitter Heron was open sourced and used by several companies. He has two decades of experience working with companies such as Teradata, Greenplum, and Juniper in their rapid growth stages building parallel databases, big data infrastructure, and networking. He co-founded Locomatix, a company that specializes in real-time streaming processing on Hadoop and Cassandra using SQL, that was acquired by Twitter. Karthik has a Ph.D. in computer science from the University of Wisconsin, Madison with a focus on big data and databases. During his college tenure several of his research projects were later spun off as a company acquired by Teradata. Karthik is the author of several publications, patents, and a popular book "Network Routing: Algorithms, Protocols and Architectures".
The number of computers in the world is still growing at an exponential rate. For instance, there are 9 Billion new micro-controllers shipped every year and that is on top of other devices such as mobile phones and servers in the cloud and on-premise. This talk takes a different perspective on this trend and argues that it makes sense to look at all these micro-controllers, devices, and servers as one big supercomputer.
The talk gives examples that illustrate how to make this supercomputer sustainable, secure, programable, and useful.
Donald Kossmann is the director of the Microsoft Research Lab in Redmond. Before joining Microsoft, Donald was a professor in the Systems Group of the Department of Computer Science at ETH Zurich for 13 years, doing research and teaching all flavors of data management systems. He was chair of ACM SIGMOD from 2013 to 2017 and served on the Board of Trustees of the VLDB Endowment from 2005 to 2011. He is a co-founder of four companies: i-TV-T AG (1998), XQRL Inc. (2002), 28msec Inc. (2006), and Teralytics AG (2010).
In a few short years, the Function-as-a-Service paradigm has revolutionized how we think about distributed event processing. Within seconds, developers can deploy and run event-driven functions in the cloud that scale on demand. CIOs love the high availability and zero administration. CFOs love the fine-grained, pay-as-you-go pricing model.
Most FaaS platforms today are biased toward (if not restricted to) simple functions, functions that only run for a short while, use limited CPU and memory, and process relatively small amounts of data. In this talk, we will get a closer look at FaaS platforms (using Apache OpenWhisk as our exemplar) to understand these trade-offs.
Functions are not applications. To build compelling FaaS applications we need to compose functions. In this talk, we will compare approaches to function composition both from a developer and from a system’s perspective. We will show how composition can dramatically expand the scope of FaaS.
Dr. Olivier Tardieu is a Principal Research Staff Member at IBM T.J. Watson, USA. He received a Ph.D. in Computer Science from Ecole des Mines de Paris, France in 2004, and joined IBM Research in 2007. His research focuses on making developers more productive with better programming models and methodologies. His passion for programming language design has driven him to explore a wide range of topics including high-performance computing, digital circuit design, log analysis, and stream processing. Today he devotes his energy to cloud computing, and in particular the serverless paradigm.
Abstract submission for research track |
|
Research and Industry paper submission |
|
Tutorial proposal submission |
|
Grand challenge solution submission |
|
Author notification research and industry track |
|
Poster, demo & doctoral symposium submission | |
Early registration | May 31st, 2019 |