Skip to content

Commit

Permalink
ogg_pager: Fix writing of large packets
Browse files Browse the repository at this point in the history
  • Loading branch information
Serial-ATA committed Apr 18, 2024
1 parent 6230b2c commit 1f6d766
Show file tree
Hide file tree
Showing 5 changed files with 53 additions and 65 deletions.
2 changes: 1 addition & 1 deletion Cargo.toml
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ lofty_attr = { path = "lofty_attr" }
# Debug logging
log = "0.4.20"
# OGG Vorbis/Opus
ogg_pager = "0.6.0"
ogg_pager = { path = "ogg_pager" }
# Key maps
paste = "1.0.14"

Expand Down
3 changes: 3 additions & 0 deletions ogg_pager/CHANGELOG.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,9 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0

## [Unreleased]

### Fixed
- When writing large packets, the size would slowly shift out of sync, causing the pages to be written incorrectly. ([issue](https://github.com/Serial-ATA/lofty-rs/issues/350)) ([PR](https://github.com/Serial-ATA/lofty-rs/pull/375))

## [0.6.0] - 2024-01-03

### Added
Expand Down
2 changes: 0 additions & 2 deletions ogg_pager/src/lib.rs
Original file line number Diff line number Diff line change
Expand Up @@ -107,8 +107,6 @@ impl Page {

/// Attempts to get a Page from a reader
///
/// Use `skip_content` to only read the header, and skip over the content.
///
/// # Errors
///
/// * [`std::io::Error`]
Expand Down
105 changes: 46 additions & 59 deletions ogg_pager/src/paginate.rs
Original file line number Diff line number Diff line change
@@ -1,8 +1,7 @@
use crate::error::Result;
use crate::{
segment_table, Page, PageHeader, CONTAINS_FIRST_PAGE_OF_BITSTREAM,
CONTAINS_LAST_PAGE_OF_BITSTREAM, CONTINUED_PACKET, MAX_WRITTEN_CONTENT_SIZE,
MAX_WRITTEN_SEGMENT_COUNT,
Page, PageHeader, CONTAINS_FIRST_PAGE_OF_BITSTREAM, CONTAINS_LAST_PAGE_OF_BITSTREAM,
CONTINUED_PACKET, MAX_WRITTEN_CONTENT_SIZE, MAX_WRITTEN_SEGMENT_COUNT,
};

use std::io::Read;
Expand All @@ -17,6 +16,7 @@ struct PaginateContext {
idx: usize,
remaining_page_size: usize,
current_packet_len: usize,
last_segment_size: u8,
}

impl PaginateContext {
Expand All @@ -36,10 +36,19 @@ impl PaginateContext {
idx: 0,
remaining_page_size: MAX_WRITTEN_CONTENT_SIZE,
current_packet_len: 0,
last_segment_size: 0,
}
}

fn flush(&mut self, content: &mut Vec<u8>, segment_table: &mut Vec<u8>) {
fn fresh_packet(&mut self, packet: &[u8]) {
self.flags.fresh_packet = true;
self.pos = 0;

self.current_packet_len = packet.len();
self.last_segment_size = (packet.len() % 255) as u8;
}

fn flush_page(&mut self, content: &mut Vec<u8>) {
let mut header = PageHeader {
start: self.pos,
header_type_flag: {
Expand All @@ -62,7 +71,7 @@ impl PaginateContext {
stream_serial: self.stream_serial,
sequence_number: self.idx as u32,
segments: Vec::new(),
// No need to calculate this yet
// Calculated later
checksum: 0,
};

Expand All @@ -71,23 +80,26 @@ impl PaginateContext {
self.pos += content_len as u64;

// Moving on to a new packet
if self.pos > self.current_packet_len as u64 {
debug_assert!(self.pos <= self.current_packet_len as u64);
if self.pos == self.current_packet_len as u64 {
self.flags.packet_spans_multiple_pages = false;
}

// We need to determine how many segments our page content takes up.
// If it takes up the remainder of the segment table for the entire packet,
// we'll just consume it as is.
let segments_occupied = if content_len >= 255 {
content_len / 255
content_len.div_ceil(255)
} else {
1
};

debug_assert!(segments_occupied <= MAX_WRITTEN_SEGMENT_COUNT);
if self.flags.packet_spans_multiple_pages {
header.segments = segment_table.drain(..segments_occupied).collect();
header.segments = vec![255; segments_occupied];
} else {
header.segments = core::mem::take(segment_table);
header.segments = vec![255; segments_occupied - 1];
header.segments.push(self.last_segment_size);
}

self.pages.push(Page {
Expand Down Expand Up @@ -137,62 +149,43 @@ where
{
let mut ctx = PaginateContext::new(abgp, stream_serial, flags);

let mut packets_iter = packets.into_iter();
let mut packet = match packets_iter.next() {
Some(packet) => packet,
// We weren't given any content to paginate
None => return Ok(ctx.pages),
};
ctx.current_packet_len = packet.len();
for packet in packets {
ctx.fresh_packet(packet);
paginate_packet(&mut ctx, packet)?;
}

if flags & CONTAINS_LAST_PAGE_OF_BITSTREAM == 0x04 {
if let Some(last) = ctx.pages.last_mut() {
last.header.header_type_flag |= CONTAINS_LAST_PAGE_OF_BITSTREAM;
}
}

let mut segments = segment_table(packet.len());
let mut page_content = Vec::new();
Ok(ctx.pages)
}

fn paginate_packet(ctx: &mut PaginateContext, packet: &[u8]) -> Result<()> {
let mut page_content = Vec::with_capacity(MAX_WRITTEN_CONTENT_SIZE);
let mut packet = packet;
loop {
if !ctx.flags.packet_spans_multiple_pages && !ctx.flags.first_page {
match packets_iter.next() {
Some(packet_) => {
packet = packet_;
segments.append(&mut segment_table(packet.len()));
ctx.current_packet_len = packet.len();
ctx.flags.fresh_packet = true;
},
None => break,
};
if packet.is_empty() {
break;
}

// We read as much of the packet as we can, given the amount of space left in the page.
// The packet may need to span multiple pages.
let bytes_read = packet
.take(ctx.remaining_page_size as u64)
.read_to_end(&mut page_content)?;
ctx.remaining_page_size -= bytes_read;

packet = &packet[bytes_read..];
// We need to indicate whether or not any packet was finished on this page.
// This is used for the absolute granule position.
if packet.is_empty() {

if bytes_read <= MAX_WRITTEN_CONTENT_SIZE && packet.is_empty() {
ctx.flags.packet_finished_on_page = true;
} else {
ctx.flags.packet_spans_multiple_pages = true;
}

// The first packet of the bitstream must have its own page, unlike any other packet.
let first_page_of_bitstream = ctx.header_flags & CONTAINS_FIRST_PAGE_OF_BITSTREAM != 0;
let first_packet_finished_on_page =
ctx.flags.first_page && first_page_of_bitstream && ctx.flags.packet_finished_on_page;

// We have a maximum of `MAX_WRITTEN_SEGMENT_COUNT` segments available per page, if we require more than
// is left in the segment table, we'll have to split the packet into multiple pages.
let segments_required = (packet.len() / MAX_WRITTEN_SEGMENT_COUNT) + 1;
let remaining_segments = MAX_WRITTEN_SEGMENT_COUNT.saturating_sub(segments.len());
ctx.flags.packet_spans_multiple_pages = segments_required > remaining_segments;

if first_packet_finished_on_page
// We've completely filled this page, we need to flush before moving on
|| (ctx.remaining_page_size == 0 || remaining_segments == 0)
// We've read all this packet has to offer
|| packet.is_empty()
{
ctx.flush(&mut page_content, &mut segments);
if ctx.remaining_page_size == 0 || packet.is_empty() {
ctx.flush_page(&mut page_content);
}

ctx.flags.first_page = false;
Expand All @@ -201,14 +194,8 @@ where

// Flush any content leftover
if !page_content.is_empty() {
ctx.flush(&mut page_content, &mut segments);
ctx.flush_page(&mut page_content);
}

if flags & CONTAINS_LAST_PAGE_OF_BITSTREAM == 0x04 {
if let Some(last) = ctx.pages.last_mut() {
last.header.header_type_flag |= CONTAINS_LAST_PAGE_OF_BITSTREAM;
}
}

Ok(ctx.pages)
Ok(())
}
6 changes: 3 additions & 3 deletions src/ogg/read.rs
Original file line number Diff line number Diff line change
Expand Up @@ -72,15 +72,15 @@ where
},
};

let comments_total_len = data.read_u32::<LittleEndian>()?;
let number_of_items = data.read_u32::<LittleEndian>()?;

let mut tag = VorbisComments {
vendor,
items: Vec::with_capacity(comments_total_len as usize),
items: Vec::with_capacity(number_of_items as usize),
pictures: Vec::new(),
};

for _ in 0..comments_total_len {
for _ in 0..number_of_items {
let comment_len = data.read_u32::<LittleEndian>()?;
if u64::from(comment_len) > len {
err!(SizeMismatch);
Expand Down

0 comments on commit 1f6d766

Please sign in to comment.