Welcome to the STREAMIND Grand Challenge!
The objective of the challenge is to design and implement a complete streaming pipeline, bridging the gap between live audio reception and structured language model inference. Participants will build their solutions on top of Juturna, an open-source Python framework developed by Meetecho for real-time AI data pipeline prototyping.
- Registration deadline: April 3, 2026
- Paper submission deadline: June 19, 2026
- Acceptance notification: July 17, 2026
To register, submit a form here.
The core task of this challenge is the implementation of a pipeline that consumes live audio streams as inputs, and produces textual summaries and keywords as outputs.
Roughly speaking, such pipeline can be organized around five main processing steps:
- Audio reception via WebRTC/RTP: a live audio stream is delivered as a Opus-encoded mono RTP stream on a dedicated port.
- Incremental ASR transcription: the incoming audio stream is incrementally transcribed by an ASR node operating on short, consecutive and partially overlapping audio chunks of a set temporal length.
- Novel chunk extraction: novel transcript text is isolated while content from previous audio chunks is discarded.
- Window aggregation: transcription chunks are accumulated into context windows of 300 seconds.
- Summarization: aggregated windows are processed to produce structured textual outputs.
- Transmission: summary objects are stored locally on the filesystem and
transmitted to a destination endpoint through
POSTrequests.
All submitted chunks for an audio source will be assigned a composite score.
Chunk
The final chunk score
The final score for an audio source is the average of all scores of its chunks.
- Juturna framework: https://github.com/meetecho/juturna
- Janus WebRTC server: https://github.com/meetecho/janus-gateway
