Commit | Line | Data |
---|---|---|
b0481c9e KS |
1 | use nihav_core::frame::*; |
2 | use nihav_core::soundcvt::*; | |
3 | ||
4 | struct AudioQueue<T> { | |
5 | start: usize, | |
6 | end: usize, | |
7 | stride: usize, | |
8 | channels: usize, | |
9 | data: Vec<T>, | |
10 | ileaved: bool, | |
11 | } | |
12 | ||
13 | impl<T:Clone+Copy+From<u8>> AudioQueue<T> { | |
14 | fn new(channels: usize, rec_size: usize, ileaved: bool) -> Self { | |
15 | Self { | |
16 | start: 0, | |
17 | end: 0, | |
18 | stride: if ileaved { rec_size * channels } else { rec_size }, | |
19 | channels, | |
20 | ileaved, | |
21 | data: vec![0.into(); rec_size * channels], | |
22 | } | |
23 | } | |
803e17a2 KS |
24 | fn get_cur_size(&self) -> usize { |
25 | let size = self.end - self.start; | |
26 | if !self.ileaved { | |
27 | size | |
28 | } else { | |
29 | size / self.channels | |
30 | } | |
31 | } | |
b0481c9e KS |
32 | fn get_cur_avail(&self) -> usize { self.stride - self.end } |
33 | fn get_potentially_avail(&self) -> usize { self.stride - self.get_cur_size() } | |
34 | fn read(&mut self, src: &NAAudioBuffer<T>) { | |
35 | let mut to_copy = src.get_length(); | |
36 | if self.ileaved { | |
37 | to_copy *= self.channels; | |
38 | } | |
39 | if self.get_cur_avail() < to_copy { | |
40 | if self.get_potentially_avail() >= to_copy { | |
41 | self.renorm(); | |
42 | } else { | |
43 | let new_len = (self.stride * 2).max(self.get_cur_size() + src.get_length()); | |
44 | let mut new_buf = vec![0.into(); new_len * self.channels]; | |
45 | let new_stride = if !self.ileaved { new_len } else { new_len * self.channels }; | |
46 | ||
47 | let old_len = self.get_cur_size(); | |
48 | let new_len = src.get_length(); | |
6f6ba7bf KS |
49 | if old_len > 0 { |
50 | for (dst, (old, new)) in new_buf.chunks_exact_mut(new_stride).zip( | |
51 | self.data.chunks_exact(self.stride).zip( | |
52 | src.get_data().chunks(src.get_stride()))) { | |
53 | dst[..old_len].copy_from_slice(&old[self.start..self.end]); | |
54 | dst[old_len..][..new_len].copy_from_slice(&new[..new_len]); | |
55 | } | |
56 | } else { | |
57 | for (dst, new) in new_buf.chunks_exact_mut(new_stride).zip( | |
58 | src.get_data().chunks(src.get_stride())) { | |
59 | dst[..new_len].copy_from_slice(&new[..new_len]); | |
60 | } | |
b0481c9e KS |
61 | } |
62 | self.data = new_buf; | |
63 | self.stride = new_stride; | |
64 | self.start = 0; | |
65 | self.end = old_len + new_len; | |
66 | return; | |
67 | } | |
68 | } | |
803e17a2 KS |
69 | match (src.get_step() != 1, self.ileaved) { |
70 | (false, false) => { | |
71 | for (dst, src) in self.data.chunks_exact_mut(self.stride).zip(src.get_data().chunks_exact(src.get_stride())) { | |
72 | dst[self.end..][..to_copy].copy_from_slice(&src[..to_copy]); | |
73 | } | |
74 | }, | |
75 | (true, false) => { | |
76 | for (i, chunk) in src.get_data().chunks_exact(src.get_step()).enumerate() { | |
77 | for (ch, &samp) in chunk.iter().enumerate() { | |
78 | self.data[self.stride * ch + self.end + i] = samp; | |
79 | } | |
80 | } | |
81 | } | |
82 | (true, true) => { | |
83 | let sdata = src.get_data(); | |
84 | self.data[self.end..][..to_copy].copy_from_slice(&sdata[..to_copy]); | |
85 | }, | |
86 | _ => unimplemented!(), | |
87 | }; | |
b0481c9e KS |
88 | self.end += to_copy; |
89 | } | |
90 | fn write(&mut self, dbuf: &mut NAAudioBuffer<T>) { | |
803e17a2 | 91 | let mut dst_len = dbuf.get_length(); |
b0481c9e | 92 | let dst_stride = dbuf.get_stride(); |
803e17a2 | 93 | let dst_step = dbuf.get_step(); |
b0481c9e KS |
94 | let dst = dbuf.get_data_mut().unwrap(); |
95 | ||
803e17a2 KS |
96 | match (self.ileaved, dst_step != 1) { |
97 | (false, false) => { | |
98 | for (dst, src) in dst.chunks_mut(dst_stride).zip(self.data.chunks_exact(self.stride)) { | |
99 | dst[..dst_len].copy_from_slice(&src[self.start..][..dst_len]); | |
100 | } | |
101 | }, | |
102 | (true, true) => { | |
103 | dst_len *= self.channels; | |
104 | dst[..dst_len].copy_from_slice(&self.data[self.start..][..dst_len]); | |
105 | }, | |
106 | _ => unimplemented!(), | |
107 | }; | |
b0481c9e KS |
108 | self.start += dst_len; |
109 | } | |
110 | fn renorm(&mut self) { | |
111 | if self.start == 0 { | |
112 | return; | |
113 | } | |
114 | ||
115 | let move_size = self.end - self.start; | |
116 | if move_size > 0 { | |
117 | for chan in self.data.chunks_exact_mut(self.stride) { | |
118 | for i in 0..move_size { | |
119 | chan[i] = chan[self.start + i]; | |
120 | } | |
121 | } | |
122 | } | |
123 | self.end -= self.start; | |
124 | self.start = 0; | |
125 | } | |
126 | } | |
127 | ||
128 | enum AudioDataType { | |
129 | U8(AudioQueue<u8>), | |
130 | I16(AudioQueue<i16>), | |
131 | I32(AudioQueue<i32>), | |
132 | F32(AudioQueue<f32>), | |
133 | Packed(AudioQueue<u8>), | |
134 | } | |
135 | ||
136 | impl AudioDataType { | |
137 | fn get_length(&self) -> usize { | |
138 | match self { | |
139 | AudioDataType::U8(ref queue) => queue.get_cur_size(), | |
140 | AudioDataType::I16(ref queue) => queue.get_cur_size(), | |
141 | AudioDataType::I32(ref queue) => queue.get_cur_size(), | |
142 | AudioDataType::F32(ref queue) => queue.get_cur_size(), | |
143 | AudioDataType::Packed(ref queue) => queue.get_cur_size(), | |
144 | } | |
145 | } | |
146 | } | |
147 | ||
148 | pub struct AudioConverter { | |
149 | queue: AudioDataType, | |
150 | dst_fmt: NAAudioInfo, | |
151 | dst_chmap: NAChannelMap, | |
152 | apts: Option<u64>, | |
6f6ba7bf | 153 | resampler: NAResample, |
b0481c9e KS |
154 | } |
155 | ||
156 | impl AudioConverter { | |
6f6ba7bf | 157 | pub fn new(sinfo: &NAAudioInfo, dinfo: &NAAudioInfo, dst_chmap: NAChannelMap) -> Self { |
b0481c9e KS |
158 | let ch = usize::from(dinfo.channels); |
159 | let size = dinfo.block_len * 2; | |
160 | let il = !dinfo.format.planar; | |
161 | let queue = match (dinfo.format.bits, dinfo.format.float, dinfo.format.signed) { | |
162 | ( 8, false, false) => AudioDataType::U8(AudioQueue::new(ch, size, il)), | |
163 | (16, false, true) => AudioDataType::I16(AudioQueue::new(ch, size, il)), | |
164 | (32, false, true) => AudioDataType::I32(AudioQueue::new(ch, size, il)), | |
165 | (32, true, _) => AudioDataType::F32(AudioQueue::new(ch, size, il)), | |
166 | _ => AudioDataType::Packed(AudioQueue::new(ch, size, il)), | |
167 | }; | |
6f6ba7bf KS |
168 | const RESAMPLE_FILTER_ORDER: usize = 16; |
169 | let resampler = NAResample::new(sinfo.sample_rate, dinfo, &dst_chmap, RESAMPLE_FILTER_ORDER); | |
b0481c9e KS |
170 | Self { |
171 | queue, | |
172 | dst_fmt: *dinfo, | |
173 | dst_chmap, | |
174 | apts: None, | |
6f6ba7bf | 175 | resampler, |
b0481c9e KS |
176 | } |
177 | } | |
178 | pub fn queue_frame(&mut self, buf: NABufferType, tinfo: NATimeInfo) -> bool { | |
6f6ba7bf | 179 | let ret = self.resampler.convert_audio_frame(&buf); |
b0481c9e KS |
180 | if let Ok(dbuf) = ret { |
181 | if self.apts.is_none() && tinfo.get_pts().is_some() { | |
182 | self.apts = tinfo.get_pts(); | |
183 | } | |
184 | match (&mut self.queue, dbuf) { | |
185 | (AudioDataType::U8(ref mut queue), NABufferType::AudioU8(ref buf)) => queue.read(buf), | |
186 | (AudioDataType::I16(ref mut queue), NABufferType::AudioI16(ref buf)) => queue.read(buf), | |
187 | (AudioDataType::I32(ref mut queue), NABufferType::AudioI32(ref buf)) => queue.read(buf), | |
188 | (AudioDataType::F32(ref mut queue), NABufferType::AudioF32(ref buf)) => queue.read(buf), | |
189 | (AudioDataType::Packed(ref mut queue), NABufferType::AudioPacked(ref buf)) => queue.read(buf), | |
190 | _ => unimplemented!(), | |
191 | }; | |
192 | true | |
193 | } else { | |
194 | false | |
195 | } | |
196 | } | |
197 | pub fn get_frame(&mut self, info: NACodecInfoRef) -> Option<NAFrame> { | |
198 | if self.queue.get_length() >= self.dst_fmt.block_len { | |
199 | if let Ok(mut abuf) = alloc_audio_buffer(self.dst_fmt, self.dst_fmt.block_len, self.dst_chmap.clone()) { | |
200 | match (&mut self.queue, &mut abuf) { | |
201 | (AudioDataType::U8(ref mut queue), NABufferType::AudioU8(ref mut buf)) => queue.write(buf), | |
202 | (AudioDataType::I16(ref mut queue), NABufferType::AudioI16(ref mut buf)) => queue.write(buf), | |
203 | (AudioDataType::I32(ref mut queue), NABufferType::AudioI32(ref mut buf)) => queue.write(buf), | |
204 | (AudioDataType::F32(ref mut queue), NABufferType::AudioF32(ref mut buf)) => queue.write(buf), | |
205 | (AudioDataType::Packed(ref mut queue), NABufferType::AudioPacked(ref mut buf)) => queue.write(buf), | |
206 | _ => unimplemented!(), | |
207 | }; | |
208 | let tinfo = NATimeInfo::new(self.apts, None, Some(self.dst_fmt.block_len as u64), 1, self.dst_fmt.sample_rate); | |
209 | if let Some(ref mut val) = self.apts { | |
210 | *val += self.dst_fmt.block_len as u64; | |
211 | } | |
212 | Some(NAFrame::new(tinfo, FrameType::I, true, info, abuf)) | |
213 | } else { | |
214 | println!(" failed to allocate audio frame"); | |
215 | None | |
216 | } | |
217 | } else { | |
218 | None | |
219 | } | |
220 | } | |
221 | } |