Papers
arxiv:2601.01554

MOSS Transcribe Diarize: Accurate Transcription with Speaker Diarization

Published on Jan 4
ยท Submitted by
Zhaoye Fei
on Jan 7
#3 Paper of the day
ยท OpenMOSS-Team OpenMOSS
Authors:
,
,
,
,
,
,
,
,
,
,
,
,
,
Zhe Xu ,
,
,
,
,

Abstract

A unified multimodal large language model for end-to-end speaker-attributed, time-stamped transcription with extended context window and strong generalization across benchmarks.

AI-generated summary

Speaker-Attributed, Time-Stamped Transcription (SATS) aims to transcribe what is said and to precisely determine the timing of each speaker, which is particularly valuable for meeting transcription. Existing SATS systems rarely adopt an end-to-end formulation and are further constrained by limited context windows, weak long-range speaker memory, and the inability to output timestamps. To address these limitations, we present MOSS Transcribe Diarize, a unified multimodal large language model that jointly performs Speaker-Attributed, Time-Stamped Transcription in an end-to-end paradigm. Trained on extensive real wild data and equipped with a 128k context window for up to 90-minute inputs, MOSS Transcribe Diarize scales well and generalizes robustly. Across comprehensive evaluations, it outperforms state-of-the-art commercial systems on multiple public and in-house benchmarks.

Community

Paper author Paper submitter

MOSS Transcribe Diarize ๐ŸŽ™๏ธ

We introduce MOSS Transcribe Diarize โ€” a unified multimodal model for Speaker-Attributed, Time-Stamped Transcription (SATS).

๐Ÿ” End-to-end SATS in a single pass (transcription + speaker attribution + timestamps)
๐Ÿง  128k context window for up to ~90-minute audio without chunking (strong long-range speaker memory)
๐ŸŒ Trained on extensive in-the-wild conversations + controllable simulated mixtures (robust to overlap/noise/domain shift)
๐Ÿ“Š Strong results on AISHELL-4 / Podcast / Movies benchmarks (best cpCER / ฮ”cp among evaluated systems)

Paper: [2601.01554] MOSS Transcribe Diarize: Accurate Transcription with Speaker Diarization
Homepage: https://mosi.cn/models/moss-transcribe-diarize
Online Demo: https://moss-transcribe-diarize-demo.mosi.cn

Verycool & useful work! Is the model going to be open source?

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2601.01554 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2601.01554 in a dataset README.md to link it from this page.

Spaces citing this paper 1

Collections including this paper 1