这个其实就是从Audio_processing.h中拿出来的。
APM should be placed in the signal chain as close to the audio hardware abstraction layer (HAL) as possible.
APM accepts only 16-bit linear PCM audio data in frames of 10 ms.Multiple channels should be interleaved.
Audio Processing instantiation and configuration:
AudioProcessing * apm = AudioProcessing :: Create (0);
apm-> level_estimator () -> Enable (true); // Enable retries estimation component
apm-> echo_cancellation () -> Enable (true); // Enable echo cancellation module
apm-> echo_cancellation () -> enable_metrics (true); //
apm-> echo_cancellation () -> enable_drift_compensation (true); // Enable clock compensation module (sound capture device clock frequency clock frequency and playback devices may be different)
apm-> gain_control () -> Enable (true); // Enable gain control module, client must enable Oh!
apm-> high_pass_filter () -> Enable (true); // high-pass filter components, DC offset and low frequency noise filtering, client must be enabled
apm-> noise_suppression () -> Enable (true); // noise suppression components, client must be enabled
apm-> voice_detection () -> Enable (true); // enable voice detection component, to detect whether there voices
apm-> voice_detection () -> set_likelihood (VoiceDetection :: kModerateLikelihood); // Set the voice detection threshold, the threshold bigger voice less likely to be ignored, some noise may be treated the same voice.
apm-> Initialize (); // Reserved internal state set by the user in all cases of re-initialization apm, to start processing a new audio stream. After creating the first stream does not necessarily need to call this method.
2.AudioProcessing workflow:
AudioProcessing is event-driven, event into the Initialize event, capturing audio event, rendering the audio event.
Initialize event:
apm-> set_sample_rate_hz (sample_rate_hz); // set the sample rate of local and remote audio stream
apm-> echo_cancellation () -> set_device_sample_rate_hz (); // set the sample rate audio equipment, we assume that the audio capture and playback device using the same sampling rate. (Must be called when the drift component is enabled)
apm-> set_num_channels (num_capture_input_channels, num_capture_output_channels);
// set the local and remote audio stream of the number of channels
Play event:
apm-> AnalyzeReverseStream (& far_frame)); // analysis of 10ms frame data far end of the audio stream, these data provide a reference for echo suppression. (Enable echo suppression when calling needs)
Capture events:
apm-> gain_control () -> set_stream_analog_level (capture_level);
apm-> set_stream_delay_ms (delay_ms + extra_delay_ms); // set the delay in milliseconds between local and remote audio streams of. This delay is the time difference and the distal end of the audio stream between the local audio streams, calculated as:
delay = (t_render – t_analyze) + (t_process – t_capture) ;
Among them
t_analyze end audio stream is time to AnalyzeReverseStream () method;
t_render is just similar to the distal end of the audio frame playback time;
t_capture local audio capture time frame;
t_process is the same time frame was given to local audio ProcessStream () method.
apm-> echo_cancellation () -> set_stream_drift_samples (drift_samples); // Set the difference between the audio device to capture and playback sampling rate. (Must be called when the drift component is enabled)
int err = apm-> ProcessStream (& near_frame); // processing audio streams, including all aspects of the deal. (Such as gain adjustment, echo cancellation, noise suppression, voice activity detection, high throughput rate without decoding Oh! Do for pcm data processing)
capture_level = apm-> gain_control () -> stream_analog_level (); // under emulation mode, you must call this method after ProcessStream, get the recommended analog value of new audio HAL.
stream_has_voice = apm-> voice_detection () -> stream_has_voice (); // detect whether there is a voice, you must call this method after ProcessStream
ns_speech_prob = apm-> noise_suppression () -> speech_probability (); // returns the internal voice priority calculated the probability of the current frame.
3.AudioProcessing release
AudioProcessing :: Destroy (apm);
apm = NULL;
另一个示例
AudioProcessing* apm = AudioProcessing::Create(0);
apm->set_sample_rate_hz(32000);
Super-wideband processing.
// Mono capture and stereo render.
apm->set_num_channels(1, 1);
apm->set_num_reverse_channels(2);
apm->high_pass_filter()->Enable(true);
apm->echo_cancellation()->enable_drift_compensation(false);
apm->echo_cancellation()->Enable(true);
apm->noise_reduction()->set_level(kHighSuppression);
apm->noise_reduction()->Enable(true);
apm->gain_control()->set_analog_level_limits(0, 255);
apm->gain_control()->set_mode(kAdaptiveAnalog);
apm->gain_control()->Enable(true);
apm->voice_detection()->Enable(true);
// Start a voice call...
// ... Render frame arrives bound for the audio HAL ...
apm->AnalyzeReverseStream(render_frame);
// ... Capture frame arrives from the audio HAL ...
// Call required set_stream_ functions.
apm->set_stream_delay_ms(delay_ms);
apm->gain_control()->set_stream_analog_level(analog_level);
apm->ProcessStream(capture_frame);
// Call required stream_ functions.
analog_level = apm->gain_control()->stream_analog_level();
has_voice = apm->stream_has_voice();
// Repeate render and capture processing for the duration of the call...
// Start a new call...
apm->Initialize();
// Close the application...
AudioProcessing::Destroy(apm);
apm = NULL;
参考:
- webrtc中APM(AudioProcessing module)的使用
一,实例化和配置 AudioProcessing* apm = AudioProcessing::Create(0); //这里的0指的是channelID,只是一个标注那个通道的表示 apm-> ...
- android studio 中移除module和恢复module
一.移除Android Studio中module 在Android Studio中想要删除某个module时,在Android Studio中选中module,右键发现没有delete,如图: An ...
- Android IOS WebRTC 音视频开发总结(八十七)-- WebRTC中丢包重传NACK实现分析
本文主要介绍WebRTC中丢包重传NACK的实现,作者:weizhenwei ,文章最早发表在编风网,微信ID:befoio 支持原创,转载必须注明出处,欢迎关注我的微信公众号blacker(微信ID ...
- Android IOS WebRTC 音视频开发总结(八十六)-- WebRTC中RTP/RTCP协议实现分析
本文主要介绍WebRTC中的RTP/RTCP协议,作者:weizhenwei ,文章最早发表在编风网,微信ID:befoio 支持原创,转载必须注明出处,欢迎关注我的微信公众号blacker(微信ID ...
- webrtc中的带宽自适应算法
转自:http://www.xuebuyuan.com/1248366.html webrtc中的带宽自适应算法分为两种: 1, 发端带宽控制, 原理是由rtcp中的丢包统计来动态的增加或减少带宽,在 ...
- Node.js中exports,module.exports以及require方法
在Node.js中,使用module.exports.f = ...与使用exports.f = ...是一样的,此时exports就是module.exports的一种简写方式.但是,需要注意的是, ...
- WebRTC中的NetEQ
NetEQ使得WebRTC语音引擎能够快速且高解析度地适应不断变化的网络环境,确保了音质优美且缓冲延迟最小,其集成了自适应抖动控制以及丢包隐藏算法. WebRTC和NetEQ概述 WebRTC Web ...
- [转载]Pytorch中nn.Linear module的理解
[转载]Pytorch中nn.Linear module的理解 本文转载并援引全文纯粹是为了构建和分类自己的知识,方便自己未来的查找,没啥其他意思. 这个模块要实现的公式是:y=xAT+*b 来源:h ...
- ULPFEC在WebRTC中的实现[转载]
一.WebRTC对抗网络丢包的两种手段 丢包重传(NACK)和前向纠错(FEC).FEC是一种前向纠错技术,发送端将负载数据加上一定的冗余纠错码一起发送,接收端根据接收到的纠错码对数据进行差错 ...
随机推荐
- python基础知识(三)
摘要:主要涉及新数据类型set集合.三元运算.深浅拷贝.函数基础.全局变量与局部变量 一.set --> 无序,不允许重复的集合 不允许重复的列表, 1,创建 s = set() 接收 ...
- 【UWP】解析GB2312、GBK编码网页乱码问题
在WebHttpRequest请求网页后,获取到的中文是乱码,类似这样: <title>˹ŵ��Ϸ���������� - ��̳������ - ˹ŵ��Ϸ����</title ...
- Linux开机自动登录(文本模式)
• Linux系统启动登录过程 以RedHat/CentOS为例,Linux系统Level3模式下从启动到登录的整个过程大致如下: 1> 加载BIOS信息:包含了CPU/显卡/内存/硬盘/网卡等 ...
- Stm32F107VCT6硬件资源
一直在用stm32系列的单片机,最近因为需要使用到以太网,所以采用了stm32f107vc这款芯片.对于存储资源一直不是非常清楚,今天专门到官网看了一下. 如下: RAM------>64kB ...
- required - HTML5里的input标签的required属性提示文字修改
input 里面增加这样的语句: <input type="text" placeholder="您的姓名" required oninvalid=&qu ...
- mac gem install nokogiri error
Gem::Ext::BuildError: ERROR: Failed to build gem native extension. /Users/angela/.rbenv/versions/1.9 ...
- 学习 opencv---(4) 分离颜色通道 && 多通道混合
上篇文章中我们讲到了使用addWeighted函数进行图像混合操作,以及将ROI和addWeighted函数结合起来使用,对指定区域进行图像混合操作. 而为了更好地观察一些图像材料的特征,有时需要对R ...
- 我的LaTeX中文文档模板
中文LaTeX处理模板 环境MiTex内核 编辑环境WinEdit 源码如下: \documentclass[a4paper,12pt]{article} \usepackage{CJK} %设定字号 ...
- iOS Salesforce SDK 小知识
Salesforce SDK 能做许多任务,因此也比较繁杂,又分了原生,js等多个调用方法. 关键点总结: SFSmartStore 中的 - (id) initWithName:(NSString* ...
- Android 通用流行框架
原文出处: http://android.jobbole.com/83028/ 1. 缓存 名称 描述 DiskLruCache Java实现基于LRU的磁盘缓存 2.图片加载 名称 描述 Andro ...