Windows and triggers - CS 591 K1: Data Stream Processing and Analytics Spring 2020Processing and Analytics Vasiliki (Vasia) Kalavri vkalavri@bu.edu Spring 2020 2/11: Windows and Triggers Vasiliki Kalavri | Boston University 2020 • Practical way to perform operations on unbounded input have a start and an end timestamp. • All built-in window assigners provide a default trigger that triggers the evaluation of a window once the (processing or event) time passes the end of the window. Long, ctx: OnTimerContext, out: Collector[OUT]) is invoked when a previously registered timer triggers. The timestamp argument gives the timestamp of the firing timer and the Collector allows emitting0 码力 | 35 页 | 444.84 KB | 1 年前3
Scalable Stream Processing - Spark Streaming and FlinkSpark automatically converts this batch-like query to a streaming execution plan. ▶ 2. Specify triggers to control when to update the results. • Each time a trigger fires, Spark checks for new data (new Spark automatically converts this batch-like query to a streaming execution plan. ▶ 2. Specify triggers to control when to update the results. • Each time a trigger fires, Spark checks for new data (new Spark automatically converts this batch-like query to a streaming execution plan. ▶ 2. Specify triggers to control when to update the results. • Each time a trigger fires, Spark checks for new data (new0 码力 | 113 页 | 1.22 MB | 1 年前3
Stream ingestion and pub/sub systems - CS 591 K1: Data Stream Processing and Analytics Spring 2020overhead, latency? • Consumer receives a notification when new data is available • how to implement triggers? • Direct messaging • Direct network communication, UDP multicast, TCP • HTTP or RPC if the0 码力 | 33 页 | 700.14 KB | 1 年前3
State management - CS 591 K1: Data Stream Processing and Analytics Spring 2020state by implementing the ListCheckpointed interface • snapshotState() is invoked when Flink triggers a checkpoint of the stateful function. • restoreState() is always invoked when the job is started0 码力 | 24 页 | 914.13 KB | 1 年前3
Exactly-once fault-tolerance in Apache Flink - CS 591 K1: Data Stream Processing and Analytics Spring 2020University 2020 45 • When a source task receives a checkpoint barrier, it pauses emitting records, triggers a checkpoint of its local state at the state backend, and broadcasts barriers to all outgoing0 码力 | 81 页 | 13.18 MB | 1 年前3
PyFlink 1.15 Documentationpyflink-docs Release release-1.15 PyFlink Nov 23, 2022 CONTENTS 1 How to build docs locally 3 1.1 Getting Started . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . pyflink-docs, Release release-1.15 PyFlink is a Python API for Apache Flink that allows you to build scalable batch and streaming workloads, such as real-time data processing pipelines, large-scale exploratory HOW TO BUILD DOCS LOCALLY 1. Install dependency requirements python3 -m pip install -r dev/requirements.txt 2. Conda install pandoc conda install pandoc 3. Build the docs python3 setup.py build_sphinx0 码力 | 36 页 | 266.77 KB | 1 年前3
PyFlink 1.16 Documentationpyflink-docs Release release-1.16 PyFlink Nov 23, 2022 CONTENTS 1 How to build docs locally 3 1.1 Getting Started . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . pyflink-docs, Release release-1.16 PyFlink is a Python API for Apache Flink that allows you to build scalable batch and streaming workloads, such as real-time data processing pipelines, large-scale exploratory HOW TO BUILD DOCS LOCALLY 1. Install dependency requirements python3 -m pip install -r dev/requirements.txt 2. Conda install pandoc conda install pandoc 3. Build the docs python3 setup.py build_sphinx0 码力 | 36 页 | 266.80 KB | 1 年前3
Course introduction - CS 591 K1: Data Stream Processing and Analytics Spring 2020and processing guarantees of streaming systems • be proficient in using Apache Flink and Kafka to build end-to-end, scalable, and reliable streaming applications • have a solid understanding of how stream Vasiliki Kalavri | Boston University 2020 Final Project You will use Apache Flink and Kafka to build a real-time monitoring and anomaly detection framework for datacenters. Your framework will: •0 码力 | 34 页 | 2.53 MB | 1 年前3
High-availability, recovery semantics, and guarantees - CS 591 K1: Data Stream Processing and Analytics Spring 2020same initial state and given the same sequence of input tuples • convergent-capable: it can re-build internal state in a way that it eventually converges to a non-failure execution output • repeatable: acknowledge reception of input tuples notify upstream of oldest logged tuples necessary to re-build current state Vasiliki Kalavri | Boston University 2020 Upstream backup Recovery time • The0 码力 | 49 页 | 2.08 MB | 1 年前3
Flow control and load shedding - CS 591 K1: Data Stream Processing and Analytics Spring 2020faults caused by high congestion. • In the presence of bursty traffic, CFC causes backpressure to build up fast and propagate along congested VCs to their sources which can be throttled. • Essentially0 码力 | 43 页 | 2.42 MB | 1 年前3
共 11 条
- 1
- 2













