X-Git-Url: https://git.nihav.org/?a=blobdiff_plain;f=videoplayer%2Fsrc%2Faudiodec.rs;h=1f5c52ecc3142b016b743b12f772346c1f5c69a2;hb=HEAD;hp=3ed639881d56c241f9fab18b6e3f2e36d81466e3;hpb=4e72c04a50ab02f6be479799d5b48161f043ebd2;p=nihav-player.git diff --git a/videoplayer/src/audiodec.rs b/videoplayer/src/audiodec.rs index 3ed6398..1f5c52e 100644 --- a/videoplayer/src/audiodec.rs +++ b/videoplayer/src/audiodec.rs @@ -78,7 +78,7 @@ impl AudioQueue { if self.end + src.len() > self.queue.len() { self.queue.resize(self.end + src.len(), 0); } - self.queue[self.end..][..src.len()].copy_from_slice(&src); + self.queue[self.end..][..src.len()].copy_from_slice(src); self.end += src.len(); self.spos = samplepos; self.set_time(); @@ -122,7 +122,7 @@ impl AudioCallback for AudioOutput { type Channel = i16; fn callback(&mut self, out: &mut [Self::Channel]) { - let mut queue = self.queue.lock().unwrap(); + let mut queue = self.queue.lock().expect("audio queue should be accessible"); let dstlen = out.len(); let copylen = queue.fill().min(dstlen); let volume = AUDIO_VOLUME.load(Ordering::Relaxed) as i32; @@ -155,20 +155,21 @@ fn dummy_audio_thread(aprecv: Receiver) -> JoinHandle<()> { type AudioPlaybackType = Option>; -fn start_audio_decoding(asystem: &AudioSubsystem, ainfo: NAAudioInfo, mut audio_dec: DecoderStuff, aprecv: Receiver) -> (AudioPlaybackType, JoinHandle<()>) { +fn start_audio_decoding(asystem: &AudioSubsystem, ainfo: NAAudioInfo, sbr_hack: bool, mut audio_dec: DecoderStuff, aprecv: Receiver) -> (AudioPlaybackType, JoinHandle<()>) { let ch = ainfo.channels.max(2); + let sample_rate = if !sbr_hack { ainfo.sample_rate } else { ainfo.sample_rate * 2 }; let desired_spec = AudioSpecDesired { - freq: Some(ainfo.sample_rate as i32), + freq: Some(sample_rate as i32), channels: Some(ch), samples: None }; let dst_info = NAAudioInfo { - sample_rate: ainfo.sample_rate, + sample_rate, channels: ch, format: SND_S16_FORMAT, block_len: 0, }; - let queue = Arc::new(Mutex::new(AudioQueue::new(ainfo.sample_rate as usize, ch as usize))); + let queue = Arc::new(Mutex::new(AudioQueue::new(sample_rate as usize, ch as usize))); let qclone = queue.clone(); let ret = asystem.open_playback(None, &desired_spec, |_spec| { AudioOutput { @@ -183,9 +184,9 @@ fn start_audio_decoding(asystem: &AudioSubsystem, ainfo: NAAudioInfo, mut audio_ let adec = if let DecoderType::Audio(ref mut dec) = audio_dec.dec { dec } else { panic!("not an audio decoder!"); }; let mut samplepos = 0usize; let dst_chmap = if dst_info.channels == 2 { - NAChannelMap::from_str("L,R").unwrap() + NAChannelMap::from_str("L,R").expect("should be able to create stereo channel map") } else { - NAChannelMap::from_str("C").unwrap() + NAChannelMap::from_str("C").expect("should be able to create single-channel map") }; ADEC_STATE.set_state(DecodingState::Normal); loop { @@ -203,17 +204,20 @@ fn start_audio_decoding(asystem: &AudioSubsystem, ainfo: NAAudioInfo, mut audio_ let buf = frm.get_buffer(); if let Some(pts) = frm.get_pts() { samplepos = NATimeInfo::ts_to_time(pts, u64::from(dst_info.sample_rate), frm.ts.tb_num, frm.ts.tb_den) as usize; + if sbr_hack { + samplepos >>= 2; + } } samplepos += buf.get_audio_length(); if let Ok(out_buf) = convert_audio_frame(&buf, &dst_info, &dst_chmap) { match out_buf { NABufferType::AudioI16(abuf) => { - let mut qdata = queue.lock().unwrap(); + let mut qdata = queue.lock().expect("audio queue should be accessible"); qdata.add(abuf.get_data(), samplepos); drop(qdata); }, NABufferType::AudioPacked(abuf) => { - let mut qdata = queue.lock().unwrap(); + let mut qdata = queue.lock().expect("audio queue should be accessible"); qdata.add_bytes(abuf.get_data(), samplepos); drop(qdata); }, @@ -226,13 +230,13 @@ fn start_audio_decoding(asystem: &AudioSubsystem, ainfo: NAAudioInfo, mut audio_ Ok(PktSendEvent::GetFrames) => {}, Ok(PktSendEvent::Flush) => { adec.flush(); - let mut qdata = queue.lock().unwrap(); + let mut qdata = queue.lock().expect("audio queue should be accessible"); qdata.flush(); ADEC_STATE.set_state(DecodingState::Waiting); }, Ok(PktSendEvent::End) => break, Ok(PktSendEvent::ImmediateEnd) => { - let mut qdata = queue.lock().unwrap(); + let mut qdata = queue.lock().expect("audio queue should be accessible"); qdata.flush(); break; }, @@ -243,7 +247,7 @@ fn start_audio_decoding(asystem: &AudioSubsystem, ainfo: NAAudioInfo, mut audio_ }; } loop { - let qdata = queue.lock().unwrap(); + let qdata = queue.lock().expect("audio queue should be accessible"); if qdata.fill() == 0 || ADEC_STATE.is_flushing() { break; } @@ -260,10 +264,10 @@ pub struct AudioControl { } impl AudioControl { - pub fn new(audio_dec: Option, ainfo: Option, asystem: &AudioSubsystem) -> Self { + pub fn new(audio_dec: Option, ainfo: Option, sbr_hack: bool, asystem: &AudioSubsystem) -> Self { let (apsend, aprecv) = std::sync::mpsc::sync_channel::(20); let (adevice, athread) = if let Some(audio_dec) = audio_dec { - start_audio_decoding(asystem, ainfo.unwrap(), audio_dec, aprecv) + start_audio_decoding(asystem, ainfo.expect("audio info should be present"), sbr_hack, audio_dec, aprecv) } else { (None, dummy_audio_thread(aprecv)) }; @@ -317,7 +321,7 @@ impl AudioControl { pub fn get_queue_size(&self) -> usize { self.aqueue.len() } pub fn try_send_audio(&mut self, evt: PktSendEvent) -> bool { - if self.aqueue.len() > 0 { + if !self.aqueue.is_empty() { self.aqueue.push(evt); false } else {