]>
Commit | Line | Data |
---|---|---|
b0481c9e KS |
1 | use nihav_core::frame::*; |
2 | use nihav_core::soundcvt::*; | |
3 | ||
4 | struct AudioQueue<T> { | |
5 | start: usize, | |
6 | end: usize, | |
7 | stride: usize, | |
8 | channels: usize, | |
9 | data: Vec<T>, | |
10 | ileaved: bool, | |
11 | } | |
12 | ||
e6ff21eb KS |
13 | fn copy_audio<T:Clone+Copy>(dst: &mut [T], dstride: usize, |
14 | src: &[T], sstride: usize, | |
15 | len: usize, channels: usize) | |
16 | { | |
17 | match (sstride == 1, dstride == 1) { | |
18 | (false, false) => { | |
19 | for (dchan, schan) in dst.chunks_mut(dstride).zip( | |
20 | src.chunks(sstride)).take(channels) { | |
21 | dchan[..len].copy_from_slice(&schan[..len]); | |
22 | } | |
23 | }, | |
24 | (false, true) => { | |
25 | for (ch, schan) in src.chunks(sstride).take(channels).enumerate() { | |
26 | for (dchunk, &samp) in dst[ch..].chunks_mut(channels).zip(schan.iter()).take(len / channels) { | |
27 | dchunk[0] = samp; | |
28 | } | |
29 | } | |
30 | }, | |
31 | (true, false) => { | |
32 | for (i, frame) in src.chunks_exact(channels).take(len).enumerate() { | |
33 | for (&samp, dchan) in frame.iter().zip(dst[i..].chunks_mut(dstride)) { | |
34 | dchan[0] = samp; | |
35 | } | |
36 | } | |
37 | }, | |
38 | (true, true) => { | |
39 | dst[..len].copy_from_slice(&src[..len]); | |
40 | }, | |
41 | } | |
42 | } | |
43 | ||
b0481c9e KS |
44 | impl<T:Clone+Copy+From<u8>> AudioQueue<T> { |
45 | fn new(channels: usize, rec_size: usize, ileaved: bool) -> Self { | |
46 | Self { | |
47 | start: 0, | |
48 | end: 0, | |
49 | stride: if ileaved { rec_size * channels } else { rec_size }, | |
50 | channels, | |
51 | ileaved, | |
52 | data: vec![0.into(); rec_size * channels], | |
53 | } | |
54 | } | |
e6ff21eb KS |
55 | fn get_cur_size(&self) -> usize { self.end - self.start } |
56 | fn get_length(&self) -> usize { | |
57 | let size = self.get_cur_size(); | |
803e17a2 KS |
58 | if !self.ileaved { |
59 | size | |
60 | } else { | |
61 | size / self.channels | |
62 | } | |
63 | } | |
b0481c9e KS |
64 | fn get_cur_avail(&self) -> usize { self.stride - self.end } |
65 | fn get_potentially_avail(&self) -> usize { self.stride - self.get_cur_size() } | |
66 | fn read(&mut self, src: &NAAudioBuffer<T>) { | |
67 | let mut to_copy = src.get_length(); | |
68 | if self.ileaved { | |
69 | to_copy *= self.channels; | |
70 | } | |
71 | if self.get_cur_avail() < to_copy { | |
72 | if self.get_potentially_avail() >= to_copy { | |
73 | self.renorm(); | |
74 | } else { | |
75 | let new_len = (self.stride * 2).max(self.get_cur_size() + src.get_length()); | |
76 | let mut new_buf = vec![0.into(); new_len * self.channels]; | |
77 | let new_stride = if !self.ileaved { new_len } else { new_len * self.channels }; | |
78 | ||
79 | let old_len = self.get_cur_size(); | |
80 | let new_len = src.get_length(); | |
6f6ba7bf | 81 | if old_len > 0 { |
e6ff21eb KS |
82 | if !self.ileaved { |
83 | for (dst, (old, new)) in new_buf.chunks_exact_mut(new_stride).zip( | |
84 | self.data.chunks_exact(self.stride).zip( | |
85 | src.get_data().chunks(src.get_stride()))) { | |
86 | dst[..old_len].copy_from_slice(&old[self.start..self.end]); | |
87 | dst[old_len..][..new_len].copy_from_slice(&new[..new_len]); | |
88 | } | |
89 | } else { | |
90 | new_buf[..old_len].copy_from_slice(&self.data[self.start..self.end]); | |
91 | copy_audio(&mut new_buf[old_len..], 1, src.get_data(), src.get_stride(), new_len, self.channels); | |
6f6ba7bf KS |
92 | } |
93 | } else { | |
e6ff21eb | 94 | copy_audio(&mut new_buf, if !self.ileaved { new_stride } else { 1 }, src.get_data(), src.get_stride(), new_len, self.channels); |
b0481c9e KS |
95 | } |
96 | self.data = new_buf; | |
97 | self.stride = new_stride; | |
98 | self.start = 0; | |
99 | self.end = old_len + new_len; | |
100 | return; | |
101 | } | |
102 | } | |
e6ff21eb KS |
103 | copy_audio(&mut self.data[self.end..], if !self.ileaved { self.stride } else { 1 }, src.get_data(), src.get_stride(), |
104 | to_copy, self.channels); | |
b0481c9e KS |
105 | self.end += to_copy; |
106 | } | |
107 | fn write(&mut self, dbuf: &mut NAAudioBuffer<T>) { | |
803e17a2 | 108 | let mut dst_len = dbuf.get_length(); |
b0481c9e KS |
109 | let dst_stride = dbuf.get_stride(); |
110 | let dst = dbuf.get_data_mut().unwrap(); | |
111 | ||
e6ff21eb KS |
112 | if dst_stride == 1 { |
113 | dst_len *= self.channels; | |
114 | } | |
115 | copy_audio(dst, dst_stride, &self.data[self.start..], if !self.ileaved { self.stride } else { 1 }, | |
116 | dst_len, self.channels); | |
b0481c9e KS |
117 | self.start += dst_len; |
118 | } | |
119 | fn renorm(&mut self) { | |
120 | if self.start == 0 { | |
121 | return; | |
122 | } | |
123 | ||
124 | let move_size = self.end - self.start; | |
125 | if move_size > 0 { | |
126 | for chan in self.data.chunks_exact_mut(self.stride) { | |
127 | for i in 0..move_size { | |
128 | chan[i] = chan[self.start + i]; | |
129 | } | |
130 | } | |
131 | } | |
132 | self.end -= self.start; | |
133 | self.start = 0; | |
134 | } | |
135 | } | |
136 | ||
137 | enum AudioDataType { | |
138 | U8(AudioQueue<u8>), | |
139 | I16(AudioQueue<i16>), | |
140 | I32(AudioQueue<i32>), | |
141 | F32(AudioQueue<f32>), | |
142 | Packed(AudioQueue<u8>), | |
143 | } | |
144 | ||
145 | impl AudioDataType { | |
146 | fn get_length(&self) -> usize { | |
147 | match self { | |
e6ff21eb KS |
148 | AudioDataType::U8(ref queue) => queue.get_length(), |
149 | AudioDataType::I16(ref queue) => queue.get_length(), | |
150 | AudioDataType::I32(ref queue) => queue.get_length(), | |
151 | AudioDataType::F32(ref queue) => queue.get_length(), | |
152 | AudioDataType::Packed(ref queue) => queue.get_length(), | |
b0481c9e KS |
153 | } |
154 | } | |
155 | } | |
156 | ||
157 | pub struct AudioConverter { | |
158 | queue: AudioDataType, | |
159 | dst_fmt: NAAudioInfo, | |
160 | dst_chmap: NAChannelMap, | |
161 | apts: Option<u64>, | |
6f6ba7bf | 162 | resampler: NAResample, |
b0481c9e KS |
163 | } |
164 | ||
165 | impl AudioConverter { | |
6f6ba7bf | 166 | pub fn new(sinfo: &NAAudioInfo, dinfo: &NAAudioInfo, dst_chmap: NAChannelMap) -> Self { |
b0481c9e KS |
167 | let ch = usize::from(dinfo.channels); |
168 | let size = dinfo.block_len * 2; | |
169 | let il = !dinfo.format.planar; | |
170 | let queue = match (dinfo.format.bits, dinfo.format.float, dinfo.format.signed) { | |
171 | ( 8, false, false) => AudioDataType::U8(AudioQueue::new(ch, size, il)), | |
172 | (16, false, true) => AudioDataType::I16(AudioQueue::new(ch, size, il)), | |
173 | (32, false, true) => AudioDataType::I32(AudioQueue::new(ch, size, il)), | |
174 | (32, true, _) => AudioDataType::F32(AudioQueue::new(ch, size, il)), | |
175 | _ => AudioDataType::Packed(AudioQueue::new(ch, size, il)), | |
176 | }; | |
6f6ba7bf KS |
177 | const RESAMPLE_FILTER_ORDER: usize = 16; |
178 | let resampler = NAResample::new(sinfo.sample_rate, dinfo, &dst_chmap, RESAMPLE_FILTER_ORDER); | |
b0481c9e KS |
179 | Self { |
180 | queue, | |
181 | dst_fmt: *dinfo, | |
182 | dst_chmap, | |
183 | apts: None, | |
6f6ba7bf | 184 | resampler, |
b0481c9e KS |
185 | } |
186 | } | |
187 | pub fn queue_frame(&mut self, buf: NABufferType, tinfo: NATimeInfo) -> bool { | |
6f6ba7bf | 188 | let ret = self.resampler.convert_audio_frame(&buf); |
b0481c9e KS |
189 | if let Ok(dbuf) = ret { |
190 | if self.apts.is_none() && tinfo.get_pts().is_some() { | |
191 | self.apts = tinfo.get_pts(); | |
192 | } | |
193 | match (&mut self.queue, dbuf) { | |
194 | (AudioDataType::U8(ref mut queue), NABufferType::AudioU8(ref buf)) => queue.read(buf), | |
195 | (AudioDataType::I16(ref mut queue), NABufferType::AudioI16(ref buf)) => queue.read(buf), | |
196 | (AudioDataType::I32(ref mut queue), NABufferType::AudioI32(ref buf)) => queue.read(buf), | |
197 | (AudioDataType::F32(ref mut queue), NABufferType::AudioF32(ref buf)) => queue.read(buf), | |
198 | (AudioDataType::Packed(ref mut queue), NABufferType::AudioPacked(ref buf)) => queue.read(buf), | |
199 | _ => unimplemented!(), | |
200 | }; | |
201 | true | |
202 | } else { | |
203 | false | |
204 | } | |
205 | } | |
206 | pub fn get_frame(&mut self, info: NACodecInfoRef) -> Option<NAFrame> { | |
207 | if self.queue.get_length() >= self.dst_fmt.block_len { | |
208 | if let Ok(mut abuf) = alloc_audio_buffer(self.dst_fmt, self.dst_fmt.block_len, self.dst_chmap.clone()) { | |
209 | match (&mut self.queue, &mut abuf) { | |
210 | (AudioDataType::U8(ref mut queue), NABufferType::AudioU8(ref mut buf)) => queue.write(buf), | |
211 | (AudioDataType::I16(ref mut queue), NABufferType::AudioI16(ref mut buf)) => queue.write(buf), | |
212 | (AudioDataType::I32(ref mut queue), NABufferType::AudioI32(ref mut buf)) => queue.write(buf), | |
213 | (AudioDataType::F32(ref mut queue), NABufferType::AudioF32(ref mut buf)) => queue.write(buf), | |
214 | (AudioDataType::Packed(ref mut queue), NABufferType::AudioPacked(ref mut buf)) => queue.write(buf), | |
215 | _ => unimplemented!(), | |
216 | }; | |
217 | let tinfo = NATimeInfo::new(self.apts, None, Some(self.dst_fmt.block_len as u64), 1, self.dst_fmt.sample_rate); | |
218 | if let Some(ref mut val) = self.apts { | |
219 | *val += self.dst_fmt.block_len as u64; | |
220 | } | |
221 | Some(NAFrame::new(tinfo, FrameType::I, true, info, abuf)) | |
222 | } else { | |
223 | println!(" failed to allocate audio frame"); | |
224 | None | |
225 | } | |
226 | } else { | |
227 | None | |
228 | } | |
229 | } | |
230 | } |