| 1 | use nihav_core::frame::*; |
| 2 | use nihav_core::soundcvt::*; |
| 3 | |
| 4 | struct AudioQueue<T> { |
| 5 | start: usize, |
| 6 | end: usize, |
| 7 | stride: usize, |
| 8 | channels: usize, |
| 9 | data: Vec<T>, |
| 10 | ileaved: bool, |
| 11 | } |
| 12 | |
| 13 | fn copy_audio<T:Clone+Copy>(dst: &mut [T], dstride: usize, |
| 14 | src: &[T], sstride: usize, |
| 15 | len: usize, channels: usize) |
| 16 | { |
| 17 | match (sstride == 1, dstride == 1) { |
| 18 | (false, false) => { |
| 19 | for (dchan, schan) in dst.chunks_mut(dstride).zip( |
| 20 | src.chunks(sstride)).take(channels) { |
| 21 | dchan[..len].copy_from_slice(&schan[..len]); |
| 22 | } |
| 23 | }, |
| 24 | (false, true) => { |
| 25 | for (ch, schan) in src.chunks(sstride).take(channels).enumerate() { |
| 26 | for (dchunk, &samp) in dst[ch..].chunks_mut(channels).zip(schan.iter()).take(len / channels) { |
| 27 | dchunk[0] = samp; |
| 28 | } |
| 29 | } |
| 30 | }, |
| 31 | (true, false) => { |
| 32 | for (i, frame) in src.chunks_exact(channels).take(len).enumerate() { |
| 33 | for (&samp, dchan) in frame.iter().zip(dst[i..].chunks_mut(dstride)) { |
| 34 | dchan[0] = samp; |
| 35 | } |
| 36 | } |
| 37 | }, |
| 38 | (true, true) => { |
| 39 | dst[..len].copy_from_slice(&src[..len]); |
| 40 | }, |
| 41 | } |
| 42 | } |
| 43 | |
| 44 | impl<T:Clone+Copy+From<u8>> AudioQueue<T> { |
| 45 | fn new(channels: usize, rec_size: usize, ileaved: bool) -> Self { |
| 46 | Self { |
| 47 | start: 0, |
| 48 | end: 0, |
| 49 | stride: if ileaved { rec_size * channels } else { rec_size }, |
| 50 | channels, |
| 51 | ileaved, |
| 52 | data: vec![0.into(); rec_size * channels], |
| 53 | } |
| 54 | } |
| 55 | fn get_cur_size(&self) -> usize { self.end - self.start } |
| 56 | fn get_length(&self) -> usize { |
| 57 | let size = self.get_cur_size(); |
| 58 | if !self.ileaved { |
| 59 | size |
| 60 | } else { |
| 61 | size / self.channels |
| 62 | } |
| 63 | } |
| 64 | fn get_cur_avail(&self) -> usize { self.stride - self.end } |
| 65 | fn get_potentially_avail(&self) -> usize { self.stride - self.get_cur_size() } |
| 66 | fn read(&mut self, src: &NAAudioBuffer<T>) { |
| 67 | let mut to_copy = src.get_length(); |
| 68 | if self.ileaved { |
| 69 | to_copy *= self.channels; |
| 70 | } |
| 71 | if self.get_cur_avail() < to_copy { |
| 72 | if self.get_potentially_avail() >= to_copy { |
| 73 | self.renorm(); |
| 74 | } else { |
| 75 | let new_len = (self.stride * 2).max(self.get_cur_size() + src.get_length()); |
| 76 | let mut new_buf = vec![0.into(); new_len * self.channels]; |
| 77 | let new_stride = if !self.ileaved { new_len } else { new_len * self.channels }; |
| 78 | |
| 79 | let old_len = self.get_cur_size(); |
| 80 | let new_len = src.get_length(); |
| 81 | if old_len > 0 { |
| 82 | if !self.ileaved { |
| 83 | for (dst, (old, new)) in new_buf.chunks_exact_mut(new_stride).zip( |
| 84 | self.data.chunks_exact(self.stride).zip( |
| 85 | src.get_data().chunks(src.get_stride()))) { |
| 86 | dst[..old_len].copy_from_slice(&old[self.start..self.end]); |
| 87 | dst[old_len..][..new_len].copy_from_slice(&new[..new_len]); |
| 88 | } |
| 89 | } else { |
| 90 | new_buf[..old_len].copy_from_slice(&self.data[self.start..self.end]); |
| 91 | copy_audio(&mut new_buf[old_len..], 1, src.get_data(), src.get_stride(), new_len, self.channels); |
| 92 | } |
| 93 | } else { |
| 94 | copy_audio(&mut new_buf, if !self.ileaved { new_stride } else { 1 }, src.get_data(), src.get_stride(), new_len, self.channels); |
| 95 | } |
| 96 | self.data = new_buf; |
| 97 | self.stride = new_stride; |
| 98 | self.start = 0; |
| 99 | self.end = old_len + new_len; |
| 100 | return; |
| 101 | } |
| 102 | } |
| 103 | copy_audio(&mut self.data[self.end..], if !self.ileaved { self.stride } else { 1 }, src.get_data(), src.get_stride(), |
| 104 | to_copy, self.channels); |
| 105 | self.end += to_copy; |
| 106 | } |
| 107 | fn write(&mut self, dbuf: &mut NAAudioBuffer<T>) { |
| 108 | let mut dst_len = dbuf.get_length(); |
| 109 | let dst_stride = dbuf.get_stride(); |
| 110 | let dst = dbuf.get_data_mut().unwrap(); |
| 111 | |
| 112 | if dst_stride == 1 { |
| 113 | dst_len *= self.channels; |
| 114 | } |
| 115 | copy_audio(dst, dst_stride, &self.data[self.start..], if !self.ileaved { self.stride } else { 1 }, |
| 116 | dst_len, self.channels); |
| 117 | self.start += dst_len; |
| 118 | } |
| 119 | fn renorm(&mut self) { |
| 120 | if self.start == 0 { |
| 121 | return; |
| 122 | } |
| 123 | |
| 124 | let move_size = self.end - self.start; |
| 125 | if move_size > 0 { |
| 126 | for chan in self.data.chunks_exact_mut(self.stride) { |
| 127 | for i in 0..move_size { |
| 128 | chan[i] = chan[self.start + i]; |
| 129 | } |
| 130 | } |
| 131 | } |
| 132 | self.end -= self.start; |
| 133 | self.start = 0; |
| 134 | } |
| 135 | } |
| 136 | |
| 137 | enum AudioDataType { |
| 138 | U8(AudioQueue<u8>), |
| 139 | I16(AudioQueue<i16>), |
| 140 | I32(AudioQueue<i32>), |
| 141 | F32(AudioQueue<f32>), |
| 142 | Packed(AudioQueue<u8>), |
| 143 | } |
| 144 | |
| 145 | impl AudioDataType { |
| 146 | fn get_length(&self) -> usize { |
| 147 | match self { |
| 148 | AudioDataType::U8(ref queue) => queue.get_length(), |
| 149 | AudioDataType::I16(ref queue) => queue.get_length(), |
| 150 | AudioDataType::I32(ref queue) => queue.get_length(), |
| 151 | AudioDataType::F32(ref queue) => queue.get_length(), |
| 152 | AudioDataType::Packed(ref queue) => queue.get_length(), |
| 153 | } |
| 154 | } |
| 155 | } |
| 156 | |
| 157 | pub struct AudioConverter { |
| 158 | queue: AudioDataType, |
| 159 | dst_fmt: NAAudioInfo, |
| 160 | dst_chmap: NAChannelMap, |
| 161 | apts: Option<u64>, |
| 162 | resampler: NAResample, |
| 163 | } |
| 164 | |
| 165 | impl AudioConverter { |
| 166 | pub fn new(sinfo: &NAAudioInfo, dinfo: &NAAudioInfo, dst_chmap: NAChannelMap) -> Self { |
| 167 | let ch = usize::from(dinfo.channels); |
| 168 | let size = dinfo.block_len * 2; |
| 169 | let il = !dinfo.format.planar; |
| 170 | let queue = match (dinfo.format.bits, dinfo.format.float, dinfo.format.signed) { |
| 171 | ( 8, false, false) => AudioDataType::U8(AudioQueue::new(ch, size, il)), |
| 172 | (16, false, true) => AudioDataType::I16(AudioQueue::new(ch, size, il)), |
| 173 | (32, false, true) => AudioDataType::I32(AudioQueue::new(ch, size, il)), |
| 174 | (32, true, _) => AudioDataType::F32(AudioQueue::new(ch, size, il)), |
| 175 | _ => AudioDataType::Packed(AudioQueue::new(ch, size, il)), |
| 176 | }; |
| 177 | const RESAMPLE_FILTER_ORDER: usize = 16; |
| 178 | let resampler = NAResample::new(sinfo.sample_rate, dinfo, &dst_chmap, RESAMPLE_FILTER_ORDER); |
| 179 | Self { |
| 180 | queue, |
| 181 | dst_fmt: *dinfo, |
| 182 | dst_chmap, |
| 183 | apts: None, |
| 184 | resampler, |
| 185 | } |
| 186 | } |
| 187 | pub fn queue_frame(&mut self, buf: NABufferType, tinfo: NATimeInfo) -> bool { |
| 188 | let ret = self.resampler.convert_audio_frame(&buf); |
| 189 | if let Ok(dbuf) = ret { |
| 190 | if self.apts.is_none() && tinfo.get_pts().is_some() { |
| 191 | self.apts = tinfo.get_pts(); |
| 192 | } |
| 193 | match (&mut self.queue, dbuf) { |
| 194 | (AudioDataType::U8(ref mut queue), NABufferType::AudioU8(ref buf)) => queue.read(buf), |
| 195 | (AudioDataType::I16(ref mut queue), NABufferType::AudioI16(ref buf)) => queue.read(buf), |
| 196 | (AudioDataType::I32(ref mut queue), NABufferType::AudioI32(ref buf)) => queue.read(buf), |
| 197 | (AudioDataType::F32(ref mut queue), NABufferType::AudioF32(ref buf)) => queue.read(buf), |
| 198 | (AudioDataType::Packed(ref mut queue), NABufferType::AudioPacked(ref buf)) => queue.read(buf), |
| 199 | _ => unimplemented!(), |
| 200 | }; |
| 201 | true |
| 202 | } else { |
| 203 | false |
| 204 | } |
| 205 | } |
| 206 | pub fn get_frame(&mut self, info: NACodecInfoRef) -> Option<NAFrame> { |
| 207 | if self.queue.get_length() >= self.dst_fmt.block_len { |
| 208 | if let Ok(mut abuf) = alloc_audio_buffer(self.dst_fmt, self.dst_fmt.block_len, self.dst_chmap.clone()) { |
| 209 | match (&mut self.queue, &mut abuf) { |
| 210 | (AudioDataType::U8(ref mut queue), NABufferType::AudioU8(ref mut buf)) => queue.write(buf), |
| 211 | (AudioDataType::I16(ref mut queue), NABufferType::AudioI16(ref mut buf)) => queue.write(buf), |
| 212 | (AudioDataType::I32(ref mut queue), NABufferType::AudioI32(ref mut buf)) => queue.write(buf), |
| 213 | (AudioDataType::F32(ref mut queue), NABufferType::AudioF32(ref mut buf)) => queue.write(buf), |
| 214 | (AudioDataType::Packed(ref mut queue), NABufferType::AudioPacked(ref mut buf)) => queue.write(buf), |
| 215 | _ => unimplemented!(), |
| 216 | }; |
| 217 | let tinfo = NATimeInfo::new(self.apts, None, Some(self.dst_fmt.block_len as u64), 1, self.dst_fmt.sample_rate); |
| 218 | if let Some(ref mut val) = self.apts { |
| 219 | *val += self.dst_fmt.block_len as u64; |
| 220 | } |
| 221 | Some(NAFrame::new(tinfo, FrameType::I, true, info, abuf)) |
| 222 | } else { |
| 223 | println!(" failed to allocate audio frame"); |
| 224 | None |
| 225 | } |
| 226 | } else { |
| 227 | None |
| 228 | } |
| 229 | } |
| 230 | } |