Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
OpenHarmony
Docs
提交
73fab5d7
D
Docs
项目概览
OpenHarmony
/
Docs
大约 2 年 前同步成功
通知
161
Star
293
Fork
28
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
D
Docs
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
73fab5d7
编写于
7月 28, 2022
作者:
O
openharmony_ci
提交者:
Gitee
7月 28, 2022
浏览文件
操作
浏览文件
下载
差异文件
!7257 更新设备描述符、焦点类型和样本格式、示例代码修改
Merge pull request !7257 from 一杯丞丞汁儿/OpenHarmony-3.2-Beta2
上级
249d6ba9
1c5277d3
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
1082 addition
and
571 deletion
+1082
-571
zh-cn/application-dev/reference/apis/js-apis-audio.md
zh-cn/application-dev/reference/apis/js-apis-audio.md
+1082
-571
未找到文件。
zh-cn/application-dev/reference/apis/js-apis-audio.md
浏览文件 @
73fab5d7
...
...
@@ -36,6 +36,24 @@ getAudioManager(): AudioManager
var audioManager = audio.getAudioManager();
```
## audio.getStreamManager<sup>9+</sup>
getStreamManager(): AudioStreamManager
获取音频流管理器实例。
**系统能力:**
SystemCapability.Multimedia.Audio.Core
**返回值:**
| 类型 | 说明 |
| -------------------------------------------------| ------------------------------- |
|
[
AudioStreamManager
](
#audiostreammanager9
)
| 返回音频流管理器实例。 |
**示例:**
```
var audioStreamManager = audio.getStreamManager();
```
## audio.createAudioRenderer<sup>8+</sup>
createAudioRenderer(options: AudioRendererOptions, callback: AsyncCallback
\<
AudioRenderer>): void
...
...
@@ -44,7 +62,7 @@ createAudioRenderer(options: AudioRendererOptions, callback: AsyncCallback\<Audi
**系统能力:**
SystemCapability.Multimedia.Audio.Renderer
**参数**
:
**参数
:
**
:
| 参数名 | 类型 | 必填 | 说明 |
| -------- | ----------------------------------------------- | ---- | ---------------- |
...
...
@@ -56,31 +74,31 @@ createAudioRenderer(options: AudioRendererOptions, callback: AsyncCallback\<Audi
```
import audio from '@ohos.multimedia.audio';
var audioStreamInfo = {
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_1,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_1,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioRendererInfo = {
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.StreamUsage.STREAM_USAGE_VOICE_COMMUNICATION,
rendererFlags: 1
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.StreamUsage.STREAM_USAGE_VOICE_COMMUNICATION,
rendererFlags: 1
}
var audioRendererOptions = {
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
}
audio.createAudioRenderer(audioRendererOptions,(err, data) => {
if (err) {
console.error(`AudioRenderer Created : Error: ${err.message}`);
}
else {
console.info('AudioRenderer Created : Success : SUCCESS');
let audioRenderer = data;
}
if (err) {
console.error(`AudioRenderer Created : Error: ${err.message}`);
}
else {
console.info('AudioRenderer Created : Success : SUCCESS');
let audioRenderer = data;
}
});
```
...
...
@@ -110,29 +128,29 @@ createAudioRenderer(options: AudioRendererOptions): Promise<AudioRenderer\>
import audio from '@ohos.multimedia.audio';
var audioStreamInfo = {
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_1,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_1,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioRendererInfo = {
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.StreamUsage.STREAM_USAGE_VOICE_COMMUNICATION,
rendererFlags: 1
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.StreamUsage.STREAM_USAGE_VOICE_COMMUNICATION,
rendererFlags: 1
}
var audioRendererOptions = {
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
}
var audioRenderer;
audio.createAudioRenderer(audioRendererOptions).then((data) => {
audioRenderer = data;
console.info('AudioFrameworkRenderLog: AudioRenderer Created : Success : Stream Type: SUCCESS');
audioRenderer = data;
console.info('AudioFrameworkRenderLog: AudioRenderer Created : Success : Stream Type: SUCCESS');
}).catch((err) => {
console.info('AudioFrameworkRenderLog: AudioRenderer Created : ERROR : '+err.message);
console.info('AudioFrameworkRenderLog: AudioRenderer Created : ERROR : '+err.message);
});
```
...
...
@@ -156,30 +174,30 @@ createAudioCapturer(options: AudioCapturerOptions, callback: AsyncCallback<Audio
```
import audio from '@ohos.multimedia.audio';
var audioStreamInfo = {
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_2,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_2,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioCapturerInfo = {
source: audio.SourceType.SOURCE_TYPE_MIC,
capturerFlags: 1
source: audio.SourceType.SOURCE_TYPE_MIC,
capturerFlags: 1
}
var audioCapturerOptions = {
streamInfo: audioStreamInfo,
capturerInfo: audioCapturerInfo
streamInfo: audioStreamInfo,
capturerInfo: audioCapturerInfo
}
audio.createAudioCapturer(audioCapturerOptions,(err, data) => {
if (err) {
console.error(`AudioCapturer Created : Error: ${err.message}`);
}
else {
console.info('AudioCapturer Created : Success : SUCCESS');
let audioCapturer = data;
}
if (err) {
console.error(`AudioCapturer Created : Error: ${err.message}`);
}
else {
console.info('AudioCapturer Created : Success : SUCCESS');
let audioCapturer = data;
}
});
```
...
...
@@ -209,28 +227,28 @@ createAudioCapturer(options: AudioCapturerOptions): Promise<AudioCapturer\>
import audio from '@ohos.multimedia.audio';
var audioStreamInfo = {
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_2,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_2,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioCapturerInfo = {
source: audio.SourceType.SOURCE_TYPE_MIC,
capturerFlags: 1
source: audio.SourceType.SOURCE_TYPE_MIC,
capturerFlags: 1
}
var audioCapturerOptions = {
streamInfo: audioStreamInfo,
capturerInfo: audioCapturerInfo
streamInfo: audioStreamInfo,
capturerInfo: audioCapturerInfo
}
var audioCapturer;
audio.createAudioCapturer(audioCapturerOptions).then((data) => {
audioCapturer = data;
console.info('AudioCapturer Created : Success : Stream Type: SUCCESS');
audioCapturer = data;
console.info('AudioCapturer Created : Success : Stream Type: SUCCESS');
}).catch((err) => {
console.info('AudioCapturer Created : ERROR : '+err.message);
console.info('AudioCapturer Created : ERROR : '+err.message);
});
```
...
...
@@ -252,7 +270,7 @@ audio.createAudioCapturer(audioCapturerOptions).then((data) => {
枚举,焦点模型。
**系统能力:**
SystemCapability.Multimedia.Audio.
InterruptMod
e
**系统能力:**
SystemCapability.Multimedia.Audio.
Cor
e
| 名称 | 默认值 | 描述 |
| ---------------------------- | ------ | ---------- |
...
...
@@ -331,13 +349,14 @@ audio.createAudioCapturer(audioCapturerOptions).then((data) => {
**系统能力:**
以下各项对应的系统能力均为SystemCapability.Multimedia.Audio.Core
| 名称 | 默认值 | 描述 |
| --------------------- | ------ | -------------------------- |
| SAMPLE_FORMAT_INVALID | -1 | 无效格式。 |
| SAMPLE_FORMAT_U8 | 0 | 无符号8位整数。 |
| SAMPLE_FORMAT_S16LE | 1 | 带符号的16位整数,小尾数。 |
| SAMPLE_FORMAT_S24LE | 2 | 带符号的24位整数,小尾数。 |
| SAMPLE_FORMAT_S32LE | 3 | 带符号的32位整数,小尾数。 |
| 名称 | 默认值 | 描述 |
| ---------------------------------- | ------ | -------------------------- |
| SAMPLE_FORMAT_INVALID | -1 | 无效格式。 |
| SAMPLE_FORMAT_U8 | 0 | 无符号8位整数。 |
| SAMPLE_FORMAT_S16LE | 1 | 带符号的16位整数,小尾数。 |
| SAMPLE_FORMAT_S24LE | 2 | 带符号的24位整数,小尾数。
<br>
由于系统限制,该采样格式仅部分设备支持,请根据实际情况使用。|
| SAMPLE_FORMAT_S32LE | 3 | 带符号的32位整数,小尾数。
<br>
由于系统限制,该采样格式仅部分设备支持,请根据实际情况使用。|
| SAMPLE_FORMAT_F32LE
<sup>
9+
</sup>
| 4 | 带符号的32位整数,小尾数。
<br>
由于系统限制,该采样格式仅部分设备支持,请根据实际情况使用。|
## AudioChannel<sup>8+</sup>
...
...
@@ -409,6 +428,17 @@ audio.createAudioCapturer(audioCapturerOptions).then((data) => {
| STREAM_USAGE_VOICE_COMMUNICATION | 2 | 语音通信。 |
| STREAM_USAGE_NOTIFICATION_RINGTONE | 6 | 通知铃声。 |
## FocusType<sup>9+</sup>
表示焦点类型的枚举。
**系统能力:**
: SystemCapability.Multimedia.Audio.Core
| 名称 | 默认值 | 描述 |
| ---------------------------------- | ------ | ------------------------------- |
| FOCUS_TYPE_RECORDING | 0 | 在录制场景使用,可打断其他音频。 |
## AudioState<sup>8+</sup>
枚举,音频状态。
...
...
@@ -576,7 +606,7 @@ audio.createAudioCapturer(audioCapturerOptions).then((data) => {
描述设备连接状态变化和设备信息。
**系统能力:**
SystemCapability.Multimedia.Audio.Device
**系统能力:**
SystemCapability.Multimedia.Audio.Device
| 名称 | 类型 | 必填 | 说明 |
| :---------------- | :------------------------------------------------ | :--- | :----------------- |
...
...
@@ -651,7 +681,7 @@ setVolume(volumeType: AudioVolumeType, volume: number, callback: AsyncCallback&l
设置指定流的音量,使用callback方式异步返回结果。
**需要权限:**
ohos.permission.ACCESS_NOTIFICATION_POLICY
,
仅设置铃声(即volumeType为AudioVolumeType.RINGTONE)在静音和非静音状态切换时需要该权限。
**需要权限:**
ohos.permission.ACCESS_NOTIFICATION_POLICY
<br/>
仅设置铃声(即volumeType为AudioVolumeType.RINGTONE)在静音和非静音状态切换时需要该权限。
**系统能力:**
SystemCapability.Multimedia.Audio.Volume
...
...
@@ -667,10 +697,10 @@ setVolume(volumeType: AudioVolumeType, volume: number, callback: AsyncCallback&l
```
audioManager.setVolume(audio.AudioVolumeType.MEDIA, 10, (err) => {
if (err) {
console.error('Failed to set the volume. ${err.message}');
return;
}
if (err) {
console.error('Failed to set the volume. ${err.message}');
return;
}
console.log('Callback invoked to indicate a successful volume setting.');
});
```
...
...
@@ -681,7 +711,7 @@ setVolume(volumeType: AudioVolumeType, volume: number): Promise<void>
设置指定流的音量,使用Promise方式异步返回结果。
**需要权限:**
ohos.permission.ACCESS_NOTIFICATION_POLICY
,
仅设置铃声(即volumeType为AudioVolumeType.RINGTONE)在静音和非静音状态切换时需要该权限。
**需要权限:**
ohos.permission.ACCESS_NOTIFICATION_POLICY
<br/>
仅设置铃声(即volumeType为AudioVolumeType.RINGTONE)在静音和非静音状态切换时需要该权限。
**系统能力:**
SystemCapability.Multimedia.Audio.Volume
...
...
@@ -702,7 +732,7 @@ setVolume(volumeType: AudioVolumeType, volume: number): Promise<void>
```
audioManager.setVolume(audio.AudioVolumeType.MEDIA, 10).then(() => {
console.log('Promise returned to indicate a successful volume setting.');
console.log('Promise returned to indicate a successful volume setting.');
});
```
...
...
@@ -725,11 +755,11 @@ getVolume(volumeType: AudioVolumeType, callback: AsyncCallback<number>): v
```
audioManager.getVolume(audio.AudioVolumeType.MEDIA, (err, value) => {
if (err) {
console.error('Failed to obtain the volume. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the volume is obtained.');
if (err) {
console.error('Failed to obtain the volume. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the volume is obtained.');
});
```
...
...
@@ -757,7 +787,7 @@ getVolume(volumeType: AudioVolumeType): Promise<number>
```
audioManager.getVolume(audio.AudioVolumeType.MEDIA).then((value) => {
console.log('Promise returned to indicate that the volume is obtained.' + value);
console.log('Promise returned to indicate that the volume is obtained.' + value);
});
```
...
...
@@ -780,11 +810,11 @@ getMinVolume(volumeType: AudioVolumeType, callback: AsyncCallback<number>)
```
audioManager.getMinVolume(audio.AudioVolumeType.MEDIA, (err, value) => {
if (err) {
console.error('Failed to obtain the minimum volume. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the minimum volume is obtained.' + value);
if (err) {
console.error('Failed to obtain the minimum volume. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the minimum volume is obtained.' + value);
});
```
...
...
@@ -812,7 +842,7 @@ getMinVolume(volumeType: AudioVolumeType): Promise<number>
```
audioManager.getMinVolume(audio.AudioVolumeType.MEDIA).then((value) => {
console.log('Promised returned to indicate that the minimum volume is obtained.' + value);
console.log('Promised returned to indicate that the minimum volume is obtained.' + value);
});
```
...
...
@@ -835,11 +865,11 @@ getMaxVolume(volumeType: AudioVolumeType, callback: AsyncCallback<number>)
```
audioManager.getMaxVolume(audio.AudioVolumeType.MEDIA, (err, value) => {
if (err) {
console.error('Failed to obtain the maximum volume. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the maximum volume is obtained.' + value);
if (err) {
console.error('Failed to obtain the maximum volume. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the maximum volume is obtained.' + value);
});
```
...
...
@@ -867,7 +897,7 @@ getMaxVolume(volumeType: AudioVolumeType): Promise<number>
```
audioManager.getMaxVolume(audio.AudioVolumeType.MEDIA).then((data) => {
console.log('Promised returned to indicate that the maximum volume is obtained.');
console.log('Promised returned to indicate that the maximum volume is obtained.');
});
```
...
...
@@ -877,7 +907,7 @@ mute(volumeType: AudioVolumeType, mute: boolean, callback: AsyncCallback<void
设置指定音量流静音,使用callback方式异步返回结果。
**需要权限:**
ohos.permission.ACCESS_NOTIFICATION_POLICY
,
仅设置铃声(即volumeType为AudioVolumeType.RINGTONE)在静音和非静音状态切换时需要该权限。
**需要权限:**
ohos.permission.ACCESS_NOTIFICATION_POLICY
<br/>
仅设置铃声(即volumeType为AudioVolumeType.RINGTONE)在静音和非静音状态切换时需要该权限。
**系统能力:**
SystemCapability.Multimedia.Audio.Volume
...
...
@@ -893,11 +923,11 @@ mute(volumeType: AudioVolumeType, mute: boolean, callback: AsyncCallback<void
```
audioManager.mute(audio.AudioVolumeType.MEDIA, true, (err) => {
if (err) {
console.error('Failed to mute the stream. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the stream is muted.');
if (err) {
console.error('Failed to mute the stream. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the stream is muted.');
});
```
...
...
@@ -907,7 +937,7 @@ mute(volumeType: AudioVolumeType, mute: boolean): Promise<void>
设置指定音量流静音,使用Promise方式异步返回结果。
**需要权限:**
ohos.permission.ACCESS_NOTIFICATION_POLICY
,
仅设置铃声(即volumeType为AudioVolumeType.RINGTONE)在静音和非静音状态切换时需要该权限。
**需要权限:**
ohos.permission.ACCESS_NOTIFICATION_POLICY
<br/>
仅设置铃声(即volumeType为AudioVolumeType.RINGTONE)在静音和非静音状态切换时需要该权限。
**系统能力:**
SystemCapability.Multimedia.Audio.Volume
...
...
@@ -929,7 +959,7 @@ mute(volumeType: AudioVolumeType, mute: boolean): Promise<void>
```
audioManager.mute(audio.AudioVolumeType.MEDIA, true).then(() => {
console.log('Promise returned to indicate that the stream is muted.');
console.log('Promise returned to indicate that the stream is muted.');
});
```
...
...
@@ -953,11 +983,11 @@ isMute(volumeType: AudioVolumeType, callback: AsyncCallback<boolean>): voi
```
audioManager.isMute(audio.AudioVolumeType.MEDIA, (err, value) => {
if (err) {
console.error('Failed to obtain the mute status. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the mute status of the stream is obtained.' + value);
if (err) {
console.error('Failed to obtain the mute status. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the mute status of the stream is obtained.' + value);
});
```
...
...
@@ -986,7 +1016,7 @@ isMute(volumeType: AudioVolumeType): Promise<boolean>
```
audioManager.isMute(audio.AudioVolumeType.MEDIA).then((value) => {
console.log('Promise returned to indicate that the mute status of the stream is obtained.' + value);
console.log('Promise returned to indicate that the mute status of the stream is obtained.' + value);
});
```
...
...
@@ -1009,11 +1039,11 @@ isActive(volumeType: AudioVolumeType, callback: AsyncCallback<boolean>): v
```
audioManager.isActive(audio.AudioVolumeType.MEDIA, (err, value) => {
if (err) {
console.error('Failed to obtain the active status of the stream. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the active status of the stream is obtained.' + value);
if (err) {
console.error('Failed to obtain the active status of the stream. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the active status of the stream is obtained.' + value);
});
```
...
...
@@ -1041,7 +1071,7 @@ isActive(volumeType: AudioVolumeType): Promise<boolean>
```
audioManager.isActive(audio.AudioVolumeType.MEDIA).then((value) => {
console.log('Promise returned to indicate that the active status of the stream is obtained.' + value);
console.log('Promise returned to indicate that the active status of the stream is obtained.' + value);
});
```
...
...
@@ -1066,11 +1096,11 @@ setRingerMode(mode: AudioRingMode, callback: AsyncCallback<void>): void
```
audioManager.setRingerMode(audio.AudioRingMode.RINGER_MODE_NORMAL, (err) => {
if (err) {
console.error('Failed to set the ringer mode. ${err.message}');
return;
}
console.log('Callback invoked to indicate a successful setting of the ringer mode.');
if (err) {
console.error('Failed to set the ringer mode. ${err.message}');
return;
}
console.log('Callback invoked to indicate a successful setting of the ringer mode.');
});
```
...
...
@@ -1100,7 +1130,7 @@ setRingerMode(mode: AudioRingMode): Promise<void>
```
audioManager.setRingerMode(audio.AudioRingMode.RINGER_MODE_NORMAL).then(() => {
console.log('Promise returned to indicate a successful setting of the ringer mode.');
console.log('Promise returned to indicate a successful setting of the ringer mode.');
});
```
...
...
@@ -1123,11 +1153,11 @@ getRingerMode(callback: AsyncCallback<AudioRingMode>): void
```
audioManager.getRingerMode((err, value) => {
if (err) {
console.error('Failed to obtain the ringer mode. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the ringer mode is obtained.' + value);
if (err) {
console.error('Failed to obtain the ringer mode. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the ringer mode is obtained.' + value);
});
```
...
...
@@ -1150,7 +1180,7 @@ getRingerMode(): Promise<AudioRingMode>
```
audioManager.getRingerMode().then((value) => {
console.log('Promise returned to indicate that the ringer mode is obtained.' + value);
console.log('Promise returned to indicate that the ringer mode is obtained.' + value);
});
```
...
...
@@ -1178,11 +1208,11 @@ setAudioParameter(key: string, value: string, callback: AsyncCallback<void>
```
audioManager.setAudioParameter('key_example', 'value_example', (err) => {
if (err) {
console.error('Failed to set the audio parameter. ${err.message}');
return;
}
console.log('Callback invoked to indicate a successful setting of the audio parameter.');
if (err) {
console.error('Failed to set the audio parameter. ${err.message}');
return;
}
console.log('Callback invoked to indicate a successful setting of the audio parameter.');
});
```
...
...
@@ -1215,7 +1245,7 @@ setAudioParameter(key: string, value: string): Promise<void>
```
audioManager.setAudioParameter('key_example', 'value_example').then(() => {
console.log('Promise returned to indicate a successful setting of the audio parameter.');
console.log('Promise returned to indicate a successful setting of the audio parameter.');
});
```
...
...
@@ -1240,11 +1270,11 @@ getAudioParameter(key: string, callback: AsyncCallback<string>): void
```
audioManager.getAudioParameter('key_example', (err, value) => {
if (err) {
console.error('Failed to obtain the value of the audio parameter. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the value of the audio parameter is obtained.' + value);
if (err) {
console.error('Failed to obtain the value of the audio parameter. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the value of the audio parameter is obtained.' + value);
});
```
...
...
@@ -1274,7 +1304,7 @@ getAudioParameter(key: string): Promise<string>
```
audioManager.getAudioParameter('key_example').then((value) => {
console.log('Promise returned to indicate that the value of the audio parameter is obtained.' + value);
console.log('Promise returned to indicate that the value of the audio parameter is obtained.' + value);
});
```
...
...
@@ -1296,11 +1326,11 @@ getDevices(deviceFlag: DeviceFlag, callback: AsyncCallback<AudioDeviceDescrip
**示例:**
```
audioManager.getDevices(audio.DeviceFlag.OUTPUT_DEVICES_FLAG, (err, value) => {
if (err) {
console.error('Failed to obtain the device list. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the device list is obtained.');
if (err) {
console.error('Failed to obtain the device list. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the device list is obtained.');
});
```
...
...
@@ -1328,7 +1358,7 @@ getDevices(deviceFlag: DeviceFlag): Promise<AudioDeviceDescriptors>
```
audioManager.getDevices(audio.DeviceFlag.OUTPUT_DEVICES_FLAG).then((data) => {
console.log('Promise returned to indicate that the device list is obtained.');
console.log('Promise returned to indicate that the device list is obtained.');
});
```
...
...
@@ -1352,11 +1382,11 @@ setDeviceActive(deviceType: ActiveDeviceType, active: boolean, callback: AsyncCa
```
audioManager.setDeviceActive(audio.ActiveDeviceType.SPEAKER, true, (err) => {
if (err) {
console.error('Failed to set the active status of the device. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the device is set to the active status.');
if (err) {
console.error('Failed to set the active status of the device. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the device is set to the active status.');
});
```
...
...
@@ -1386,7 +1416,7 @@ setDeviceActive(deviceType: ActiveDeviceType, active: boolean): Promise<void&
```
audioManager.setDeviceActive(audio.ActiveDeviceType.SPEAKER, true).then(() => {
console.log('Promise returned to indicate that the device is set to the active status.');
console.log('Promise returned to indicate that the device is set to the active status.');
});
```
...
...
@@ -1409,11 +1439,11 @@ isDeviceActive(deviceType: ActiveDeviceType, callback: AsyncCallback<boolean&
```
audioManager.isDeviceActive(audio.ActiveDeviceType.SPEAKER, (err, value) => {
if (err) {
console.error('Failed to obtain the active status of the device. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the active status of the device is obtained.');
if (err) {
console.error('Failed to obtain the active status of the device. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the active status of the device is obtained.');
});
```
...
...
@@ -1442,7 +1472,7 @@ isDeviceActive(deviceType: ActiveDeviceType): Promise<boolean>
```
audioManager.isDeviceActive(audio.ActiveDeviceType.SPEAKER).then((value) => {
console.log('Promise returned to indicate that the active status of the device is obtained.' + value);
console.log('Promise returned to indicate that the active status of the device is obtained.' + value);
});
```
...
...
@@ -1467,11 +1497,11 @@ setMicrophoneMute(mute: boolean, callback: AsyncCallback<void>): void
```
audioManager.setMicrophoneMute(true, (err) => {
if (err) {
console.error('Failed to mute the microphone. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the microphone is muted.');
if (err) {
console.error('Failed to mute the microphone. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the microphone is muted.');
});
```
...
...
@@ -1501,7 +1531,7 @@ setMicrophoneMute(mute: boolean): Promise<void>
```
audioManager.setMicrophoneMute(true).then(() => {
console.log('Promise returned to indicate that the microphone is muted.');
console.log('Promise returned to indicate that the microphone is muted.');
});
```
...
...
@@ -1525,11 +1555,11 @@ isMicrophoneMute(callback: AsyncCallback<boolean>): void
```
audioManager.isMicrophoneMute((err, value) => {
if (err) {
console.error('Failed to obtain the mute status of the microphone. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the mute status of the microphone is obtained.' + value);
if (err) {
console.error('Failed to obtain the mute status of the microphone. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the mute status of the microphone is obtained.' + value);
});
```
...
...
@@ -1554,7 +1584,7 @@ isMicrophoneMute(): Promise<boolean>
```
audioManager.isMicrophoneMute().then((value) => {
console.log('Promise returned to indicate that the mute status of the microphone is obtained.', + value);
console.log('Promise returned to indicate that the mute status of the microphone is obtained.', + value);
});
```
...
...
@@ -1566,6 +1596,8 @@ on(type: 'volumeChange', callback: Callback\<VolumeEvent>): void
此接口为系统接口,三方应用不支持调用。
目前此订阅接口在单进程多AudioManager实例的使用场景下,仅最后一个实例的订阅生效,其他实例的订阅会被覆盖(即使最后一个实例没有进行订阅),因此推荐使用单一AudioManager实例进行开发。
**系统能力:**
SystemCapability.Multimedia.Audio.Volume
**参数:**
...
...
@@ -1579,9 +1611,9 @@ on(type: 'volumeChange', callback: Callback\<VolumeEvent>): void
```
audioManager.on('volumeChange', (volumeEvent) => {
console.log('VolumeType of stream: ' + volumeEvent.volumeType);
console.log('Volume level: ' + volumeEvent.volume);
console.log('Whether to updateUI: ' + volumeEvent.updateUi);
console.log('VolumeType of stream: ' + volumeEvent.volumeType);
console.log('Volume level: ' + volumeEvent.volume);
console.log('Whether to updateUI: ' + volumeEvent.updateUi);
});
```
...
...
@@ -1606,7 +1638,7 @@ on(type: 'ringerModeChange', callback: Callback\<AudioRingMode>): void
```
audioManager.on('ringerModeChange', (ringerMode) => {
console.log('Updated ringermode: ' + ringerMode);
console.log('Updated ringermode: ' + ringerMode);
});
```
...
...
@@ -1629,10 +1661,10 @@ on(type: 'deviceChange', callback: Callback<DeviceChangeAction\>): void
```
audioManager.on('deviceChange', (deviceChanged) => {
console.info("device change type : " + deviceChanged.type);
console.info("device descriptor size : " + deviceChanged.deviceDescriptors.length);
console.info("device change descriptor : " + deviceChanged.deviceDescriptors[0].deviceRole);
console.info("device change descriptor : " + deviceChanged.deviceDescriptors[0].deviceType);
console.info("device change type : " + deviceChanged.type);
console.info("device descriptor size : " + deviceChanged.deviceDescriptors.length);
console.info("device change descriptor : " + deviceChanged.deviceDescriptors[0].deviceRole);
console.info("device change descriptor : " + deviceChanged.deviceDescriptors[0].deviceType);
});
```
...
...
@@ -1655,7 +1687,7 @@ off(type: 'deviceChange', callback?: Callback<DeviceChangeAction\>): void
```
audioManager.off('deviceChange', (deviceChanged) => {
console.log("Should be no callback.");
console.log("Should be no callback.");
});
```
...
...
@@ -1679,19 +1711,19 @@ on(type: 'interrupt', interrupt: AudioInterrupt, callback: Callback\<InterruptAc
```
var interAudioInterrupt = {
streamUsage:2,
contentType:0,
pauseWhenDucked:true
streamUsage:2,
contentType:0,
pauseWhenDucked:true
};
audioManager.on('interrupt', interAudioInterrupt, (InterruptAction) => {
if (InterruptAction.actionType === 0) {
console.log("An event to gain the audio focus starts.");
console.log("Focus gain event:" + JSON.stringify(InterruptAction));
}
if (InterruptAction.actionType === 1) {
console.log("An audio interruption event starts.");
console.log("Audio interruption event:" + JSON.stringify(InterruptAction));
}
if (InterruptAction.actionType === 0) {
console.log("An event to gain the audio focus starts.");
console.log("Focus gain event:" + JSON.stringify(InterruptAction));
}
if (InterruptAction.actionType === 1) {
console.log("An audio interruption event starts.");
console.log("Audio interruption event:" + JSON.stringify(InterruptAction));
}
});
```
...
...
@@ -1715,15 +1747,15 @@ off(type: 'interrupt', interrupt: AudioInterrupt, callback?: Callback\<Interrupt
```
var interAudioInterrupt = {
streamUsage:2,
contentType:0,
pauseWhenDucked:true
streamUsage:2,
contentType:0,
pauseWhenDucked:true
};
audioManager.off('interrupt', interAudioInterrupt, (InterruptAction) => {
if (InterruptAction.actionType === 0) {
console.log("An event to release the audio focus starts.");
console.log("Focus release event:" + JSON.stringify(InterruptAction));
}
if (InterruptAction.actionType === 0) {
console.log("An event to release the audio focus starts.");
console.log("Focus release event:" + JSON.stringify(InterruptAction));
}
});
```
...
...
@@ -1748,11 +1780,11 @@ setAudioScene\(scene: AudioScene, callback: AsyncCallback<void\>\): void
```
audioManager.setAudioScene(audio.AudioScene.AUDIO_SCENE_PHONE_CALL, (err) => {
if (err) {
console.error('Failed to set the audio scene mode. ${err.message}');
return;
}
console.log('Callback invoked to indicate a successful setting of the audio scene mode.');
if (err) {
console.error('Failed to set the audio scene mode. ${err.message}');
return;
}
console.log('Callback invoked to indicate a successful setting of the audio scene mode.');
});
```
...
...
@@ -1782,9 +1814,9 @@ setAudioScene\(scene: AudioScene\): Promise<void\>
```
audioManager.setAudioScene(audio.AudioScene.AUDIO_SCENE_PHONE_CALL).then(() => {
console.log('Promise returned to indicate a successful setting of the audio scene mode.');
console.log('Promise returned to indicate a successful setting of the audio scene mode.');
}).catch ((err) => {
console.log('Failed to set the audio scene mode');
console.log('Failed to set the audio scene mode');
});
```
...
...
@@ -1806,11 +1838,11 @@ getAudioScene\(callback: AsyncCallback<AudioScene\>\): void
```
audioManager.getAudioScene((err, value) => {
if (err) {
console.error('Failed to obtain the audio scene mode. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the audio scene mode is obtained.' + value);
if (err) {
console.error('Failed to obtain the audio scene mode. ${err.message}');
return;
}
console.log('Callback invoked to indicate that the audio scene mode is obtained.' + value);
});
```
...
...
@@ -1833,9 +1865,446 @@ getAudioScene\(\): Promise<AudioScene\>
```
audioManager.getAudioScene().then((value) => {
console.log('Promise returned to indicate that the audio scene mode is obtained.' + value);
console.log('Promise returned to indicate that the audio scene mode is obtained.' + value);
}).catch ((err) => {
console.log('Failed to obtain the audio scene mode');
console.log('Failed to obtain the audio scene mode');
});
```
## AudioStreamManager<sup>9+</sup>
管理音频流。在使用AudioStreamManager的API前,需要使用
[
getStreamManager
](
#audiogetstreammanager9
)
获取AudioStreamManager实例。
### getCurrentAudioRendererInfoArray<sup>9+</sup>
getCurrentAudioRendererInfoArray(callback: AsyncCallback
<
AudioRendererChangeInfoArray
>
): void
获取当前音频渲染器的信息。使用callback异步回调。
**系统能力**
: SystemCapability.Multimedia.Audio.Renderer
**参数:**
| 名称 | 类型 | 必填 | 说明 |
| -------- | ----------------------------------- | -------- | --------------------------- |
| callback | AsyncCallback
<
[
AudioRendererChangeInfoArray
](
#audiorendererchangeinfoarray9
)
>
| 是 | 回调函数,返回当前音频渲染器的信息。 |
**示例:**
```
audioStreamManager.getCurrentAudioRendererInfoArray(async (err, AudioRendererChangeInfoArray) => {
console.info('[GET_RENDERER_STATE_1_CALLBACK] **** Get Callback Called ****');
if (err) {
console.log('getCurrentAudioRendererInfoArray :ERROR: '+err.message);
resultFlag = false;
}
else {
if (AudioRendererChangeInfoArray !=null) {
for (let i=0;i<AudioRendererChangeInfoArray.length;i++) {
AudioRendererChangeInfo = AudioRendererChangeInfoArray[i];
console.info('StreamId for '+i+' is:'+AudioRendererChangeInfo.streamId);
console.info('ClientUid for '+i+' is:'+AudioRendererChangeInfo.clientUid);
console.info('Content '+i+' is:'+AudioRendererChangeInfo.rendererInfo.content);
console.info('Stream'+i+' is:'+AudioRendererChangeInfo.rendererInfo.usage);
console.info('Flag'+i+' is:'+AudioRendererChangeInfo.rendererInfo.rendererFlags);
console.info('State for '+i+' is:'+AudioRendererChangeInfo.rendererState);
var devDescriptor = AudioRendererChangeInfo.deviceDescriptors;
for (let j=0;j<AudioRendererChangeInfo.deviceDescriptors.length; j++) {
console.info('Id:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].id);
console.info('Type:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].deviceType);
console.info('Role:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].deviceRole);
console.info('Name:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].name);
console.info('Address:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].address);
console.info('SampleRates:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].sampleRates[0]);
console.info('ChannelCount'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].channelCounts[0]);
console.info('ChannelMask:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].channelMasks);
}
}
}
}
});
```
### getCurrentAudioRendererInfoArray<sup>9+</sup>
getCurrentAudioRendererInfoArray(): Promise
<
AudioRendererChangeInfoArray
>
获取当前音频渲染器的信息。使用Promise异步回调。
**系统能力:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
| 类型 | 说明 |
| ---------------------------------------------------------------------------------| --------------------------------------- |
| Promise
<
[
AudioRendererChangeInfoArray
](
#audiorendererchangeinfoarray9
)
>
| Promise对象,返回当前音频渲染器信息。 |
**示例:**
```
await audioStreamManager.getCurrentAudioRendererInfoArray().then( function (AudioRendererChangeInfoArray) {
console.info('[GET_RENDERER_STATE_3_PROMISE] ######### Get Promise is called ##########');
if (AudioRendererChangeInfoArray!=null) {
for (let i=0;i<AudioRendererChangeInfoArray.length;i++) {
AudioRendererChangeInfo = AudioRendererChangeInfoArray[i];
console.info('StreamId for '+i+' is:'+AudioRendererChangeInfo.streamId);
console.info('ClientUid for '+i+' is:'+AudioRendererChangeInfo.clientUid);
console.info('Content '+i+' is:'+AudioRendererChangeInfo.rendererInfo.content);
console.info('Stream'+i+' is:'+AudioRendererChangeInfo.rendererInfo.usage);
console.info('Flag'+i+' is:'+AudioRendererChangeInfo.rendererInfo.rendererFlags);
console.info('State for '+i+' is:'+AudioRendererChangeInfo.rendererState);
var devDescriptor = AudioRendererChangeInfo.deviceDescriptors;
for (let j=0;j<AudioRendererChangeInfo.deviceDescriptors.length; j++) {
console.info('Id:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].id);
console.info('Type:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].deviceType);
console.info('Role:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].deviceRole);
console.info('Name:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].name);
console.info('Address:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].address);
console.info('SampleRates:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].sampleRates[0]);
console.info('ChannelCounts'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].channelCounts[0]);
console.info('ChannnelMask:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].channelMasks);
}
}
}
}).catch((err) => {
console.log('getCurrentAudioRendererInfoArray :ERROR: '+err.message);
resultFlag = false;
});
```
### getCurrentAudioCapturerInfoArray<sup>9+</sup>
getCurrentAudioCapturerInfoArray(callback: AsyncCallback
<
AudioCapturerChangeInfoArray
>
): void
获取当前音频采集器的信息。使用callback异步回调。
**系统能力:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
| 名称 | 类型 | 必填 | 说明 |
| ---------- | ----------------------------------- | --------- | -------------------------------------------------------- |
| callback | AsyncCallback
<
[
AudioCapturerChangeInfoArray
](
#audiocapturerchangeinfoarray9
)
>
| 是 | 回调函数,返回当前音频采集器的信息。 |
**示例:**
```
audioStreamManager.getCurrentAudioCapturerInfoArray(async (err, AudioCapturerChangeInfoArray) => {
console.info('[GET_CAPTURER_STATE_1_CALLBACK] **** Get Callback Called ****');
if (err) {
console.log('getCurrentAudioCapturerInfoArray :ERROR: '+err.message);
resultFlag = false;
}
else {
if (AudioCapturerChangeInfoArray !=null) {
for (let i=0;i<AudioCapturerChangeInfoArray.length;i++) {
console.info('StreamId for '+i+'is:'+AudioCapturerChangeInfoArray[i].streamId);
console.info('ClientUid for '+i+'is:'+AudioCapturerChangeInfoArray[i].clientUid);
console.info('Source for '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerInfo.source);
console.info('Flag '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerInfo.capturerFlags);
console.info('State for '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerState);
var devDescriptor = AudioCapturerChangeInfoArray[i].deviceDescriptors;
for (let j=0;j<AudioCapturerChangeInfoArray[i].deviceDescriptors.length; j++) {
console.info('Id:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].id);
console.info('Type:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].deviceType);
console.info('Role:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].deviceRole);
console.info('Name:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].name);
console.info('Address:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].address);
console.info('SampleRates:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].sampleRates[0]);
console.info('ChannelCounts'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].channelCounts[0]);
console.info('ChannelMask:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].channelMasks);
}
}
}
}
});
```
### getCurrentAudioCapturerInfoArray<sup>9+</sup>
getCurrentAudioCapturerInfoArray(): Promise
<
AudioCapturerChangeInfoArray
>
获取当前音频采集器的信息。使用Promise异步回调。
**系统能力:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
| 类型 | 说明 |
| -----------------------------------------------------------------------------| ----------------------------------- |
| Promise
<
[
AudioCapturerChangeInfoArray
](
#audiocapturerchangeinfoarray9
)
>
| Promise对象,返回当前音频渲染器信息。 |
**示例:**
```
await audioStreamManagerCB.getCurrentAudioCapturerInfoArray().then( function (AudioCapturerChangeInfoArray) {
console.info('AFCapturerChangeLog: [GET_CAP_STA_1_PR] **** Get Promise Called ****');
if (AudioCapturerChangeInfoArray!=null) {
for (let i=0;i<AudioCapturerChangeInfoArray.length;i++) {
console.info('StreamId for '+i+'is:'+AudioCapturerChangeInfoArray[i].streamId);
console.info('ClientUid for '+i+'is:'+AudioCapturerChangeInfoArray[i].clientUid);
console.info('Source for '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerInfo.source);
console.info('Flag '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerInfo.capturerFlags);
console.info('State for '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerState);
var devDescriptor = AudioCapturerChangeInfoArray[i].deviceDescriptors;
for (let j=0;j<AudioCapturerChangeInfoArray[i].deviceDescriptors.length; j++) {
console.info('Id:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].id);
console.info('Type:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].deviceType);
console.info('Role:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].deviceRole);
console.info('Name:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].name)
console.info('Address:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].address);
console.info('SampleRates:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].sampleRates[0]);
console.info('ChannelCounts'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].channelCounts[0]);
console.info('ChannelMask:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].channelMasks);
}
}
}
}).catch((err) => {
console.log('getCurrentAudioCapturerInfoArray :ERROR: '+err.message);
resultFlag = false;
});
```
### on('audioRendererChange')<sup>9+</sup>
on(type: "audioRendererChange", callback: Callback
<
AudioRendererChangeInfoArray
>
): void
监听音频渲染器更改事件。
**系统能力:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
| 名称 | 类型 | 必填 | 说明 |
| -------- | ---------- | --------- | ------------------------------------------------------------------------ |
| type | string | 是 | 事件类型,支持的事件
`'audioRendererChange'`
:当音频渲染器发生更改时触发。 |
| callback | Callback
<
[
AudioRendererChangeInfoArray
](
#audiorendererchangeinfoarry9
)
>
| 是 | 回调函数。 |
**示例:**
```
audioStreamManagerCB.on('audioRendererChange', (AudioRendererChangeInfoArray) => {
for (let i=0;i<AudioRendererChangeInfoArray.length;i++) {
AudioRendererChangeInfo = AudioRendererChangeInfoArray[i];
console.info('## RendererChange on is called for '+i+' ##');
console.info('StreamId for '+i+' is:'+AudioRendererChangeInfo.streamId);
console.info('ClientUid for '+i+' is:'+AudioRendererChangeInfo.clientUid);
console.info('Content for '+i+' is:'+AudioRendererChangeInfo.rendererInfo.content);
console.info('Stream for '+i+' is:'+AudioRendererChangeInfo.rendererInfo.usage);
console.info('Flag '+i+' is:'+AudioRendererChangeInfo.rendererInfo.rendererFlags);
console.info('State for '+i+' is:'+AudioRendererChangeInfo.rendererState);
var devDescriptor = AudioRendererChangeInfo.deviceDescriptors;
for (let j=0;j<AudioRendererChangeInfo.deviceDescriptors.length; j++) {
console.info('Id:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].id);
console.info('Type:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].deviceType);
console.info('Role:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].deviceRole);
console.info('Name:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].name);
console.info('Address:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].address);
console.info('SampleRates:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].sampleRates[0]);
console.info('ChannelCounts'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].channelCounts[0]);
console.info('ChannelMask:'+i+':'+AudioRendererChangeInfo.deviceDescriptors[j].channelMasks);
}
}
});
```
### off('audioRendererChange')<sup>9+</sup>
off(type: "audioRendererChange");
监听音频渲染器更改事件。
**系统能力:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
| 名称 | 类型 | 必填 | 说明 |
| -------- | ------- | ---- | ---------------- |
| type | string | 是 | 事件类型,支持的事件
`'audioRendererChange'`
:音频渲染器更改事件。 |
**示例:**
```
audioStreamManagerCB.off('audioRendererChange');
console.info('[RENDERER-CHANGE-ON-001] ######### RendererChange Off is called #########');
```
### on('audioCapturerChange')<sup>9+</sup>
on(type: "audioCapturerChange", callback: Callback
<
AudioCapturerChangeInfoArray
>
): void
监听音频捕获器更改事件。
**系统能力:**
SystemCapability.Multimedia.Audio.Capturer
**参数:**
| 名称 | 类型 | 必填 | 说明 |
| -------- | ------- | --------- | ------------------------------------------------------------------- ---- |
| type | string | 是 | 事件类型,支持的事件
`'audioCapturerChange'`
:当音频采集器发生更改时触发。 |
| callback | Callback
<
[
AudioCapturerChangeInfoArray
](
#audiocapturerchangeinfoarry9
)
>
| 是 | 回调函数。 |
**示例:**
```
audioStreamManager.on('audioCapturerChange', (AudioCapturerChangeInfoArray) => {
for (let i=0;i<AudioCapturerChangeInfoArray.length;i++) {
console.info(' ## CapChange on is called for element '+i+' ##');
console.info('StreamId for '+i+'is:'+AudioCapturerChangeInfoArray[i].streamId);
console.info('ClientUid for '+i+'is:'+AudioCapturerChangeInfoArray[i].clientUid);
console.info('Source for '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerInfo.source);
console.info('Flag '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerInfo.capturerFlags);
console.info('State for '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerState);
for (let j=0;j<AudioCapturerChangeInfoArray[i].deviceDescriptors.length; j++) {
console.info('Id:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].id);
console.info('Type:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].deviceType);
console.info('Role:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].deviceRole);
console.info('Name:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].name);
console.info('Address:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].address);
console.info('SampleRates:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].sampleRates[0]);
console.info('ChannelCounts'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].channelCounts[0]);
console.info('ChannelMask:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].channelMasks);
}
}
});
```
### off('audioCapturerChange')<sup>9+</sup>
off(type: "audioCapturerChange");
监听音频捕获器更改事件。
**系统能力:**
SystemCapability.Multimedia.Audio.Capturer
**参数:**
| 名称 | 类型 | 必填 | 说明 |
| -------- | -------- | --- | ------------------------------------------------------------- |
| type | string |是 | 事件类型,支持的事件
`'audioCapturerChange'`
:音频采集器更改事件。 |
**示例:**
```
audioStreamManager.off('audioCapturerChange');
console.info('[GET_CAPTURER_STATE_2_PROMISE] ######### CapturerChange Off is called #########');
```
## AudioRendererChangeInfo<sup>9+</sup>
描述音频渲染器更改信息。
**系统能力:**
以下各项对应的系统能力均为SystemCapability.Multimedia.Audio.Renderer
| 名称 | 类型 | 可读 | 可写 | 说明 |
| -------------------| ----------------------------------------- | ---- | ---- | ---------------------------- |
| streamId | number | 是 | 否 | 音频流唯一id。 |
| clientUid | number | 是 | 否 | 音频渲染器客户端应用程序的Uid。
<br/>
此接口为系统接口,三方应用不支持调用。 |
| rendererInfo |
[
AudioRendererInfo
](
#audiorendererinfo8
)
| 是 | 否 | 音频渲染器信息。 |
| rendererState |
[
AudioState
](
#audiostate
)
| 是 | 否 | 音频状态。
<br/>
此接口为系统接口,三方应用不支持调用。|
## AudioRendererChangeInfoArray<sup>9+</sup>
AudioRenderChangeInfo数组,只读。
**系统能力:**
: SystemCapability.Multimedia.Audio.Renderer
**示例:**
```
import audio from '@ohos.multimedia.audio';
var audioStreamManager;
var audioStreamManagerCB;
var Tag = "AFCapLog : ";
await audioManager.getStreamManager().then(async function (data) {
audioStreamManager = data;
console.info(Tag+'Get AudioStream Manager : Success ');
}).catch((err) => {
console.info(Tag+'Get AudioStream Manager : ERROR : '+err.message);
});
audioManager.getStreamManager((err, data) => {
if (err) {
console.error(Tag+'Get AudioStream Manager : ERROR : '+err.message);
}
else {
audioStreamManagerCB = data;
console.info(Tag+'Get AudioStream Manager : Success ');
}
});
audioStreamManagerCB.on('audioRendererChange', (AudioRendererChangeInfoArray) => {
for (let i=0;i<AudioRendererChangeInfoArray.length;i++) {
console.info(Tag+'## RendererChange on is called for '+i+' ##');
console.info(Tag+'StreamId for '+i+' is:'+AudioRendererChangeInfoArray[i].streamId);
console.info(Tag+'ClientUid for '+i+' is:'+AudioRendererChangeInfoArray[i].clientUid);
console.info(Tag+'Content for '+i+' is:'+AudioRendererChangeInfoArray[i].rendererInfo.content);
console.info(Tag+'Stream for '+i+' is:'+AudioRendererChangeInfoArray[i].rendererInfo.usage);
console.info(Tag+'Flag '+i+' is:'+AudioRendererChangeInfoArray[i].rendererInfo.rendererFlags);
console.info(Tag+'State for '+i+' is:'+AudioRendererChangeInfoArray[i].rendererState);
var devDescriptor = AudioRendererChangeInfoArray[i].deviceDescriptors;
for (let j=0;j<AudioRendererChangeInfoArray[i].deviceDescriptors.length; j++) {
console.info(Tag+'Id:'+i+':'+AudioRendererChangeInfoArray[i].deviceDescriptors[j].id);
console.info(Tag+'Type:'+i+':'+AudioRendererChangeInfoArray[i].deviceDescriptors[j].deviceType);
console.info(Tag+'Role:'+i+':'+AudioRendererChangeInfoArray[i].deviceDescriptors[j].deviceRole);
console.info(Tag+'Name:'+i+':'+AudioRendererChangeInfoArray[i].deviceDescriptors[j].name);
console.info(Tag+'Addr:'+i+':'+AudioRendererChangeInfoArray[i].deviceDescriptors[j].address);
console.info(Tag+'SR:'+i+':'+AudioRendererChangeInfoArray[i].deviceDescriptors[j].sampleRates[0]);
console.info(Tag+'C'+i+':'+AudioRendererChangeInfoArray[i].deviceDescriptors[j].channelCounts[0]);
console.info(Tag+'CM:'+i+':'+AudioRendererChangeInfoArray[i].deviceDescriptors[j].channelMasks);
}
if (AudioRendererChangeInfoArray[i].rendererState == 1 && devDescriptor != null) {
resultFlag = true;
console.info(Tag+'[RENDERER-CHANGE-ON-001] ResultFlag for '+i+' is:'+resultFlag);
}
}
});
```
## AudioCapturerChangeInfo<sup>9+</sup>
描述音频捕获器更改信息。
**系统能力:**
以下各项对应的系统能力均为SystemCapability.Multimedia.Audio.Capturer
| 名称 | 类型 | 可读 | 可写 | 说明 |
| -------------------| ----------------------------------------- | ---- | ---- | ---------------------------- |
| streamId | number | 是 | 否 | 音频流唯一id。 |
| clientUid | number | 是 | 否 | 音频渲染器客户端应用程序的Uid。
<br/>
此接口为系统接口,三方应用不支持调用。 |
| capturerInfo |
[
AudioCapturerInfo
](
#audiocaptureinfo8
)
| 是 | 否 | 音频渲染器信息。 |
| capturerState |
[
AudioState
](
#audiostate
)
| 是 | 否 | 音频状态。
<br/>
此接口为系统接口,三方应用不支持调用。|
## AudioCapturerChangeInfoArray<sup>9+</sup>
AudioCapturerChangeInfo数组,只读。
**系统能力:**
SystemCapability.Multimedia.Audio.Capturer
**示例:**
```
import audio from '@ohos.multimedia.audio';
var Tag = "AFCapLog : ";
const audioManager = audio.getAudioManager();
audioStreamManager.on('audioCapturerChange', (AudioCapturerChangeInfoArray) => {
for (let i=0;i<AudioCapturerChangeInfoArray.length;i++) {
console.info(Tag+' ## CapChange on is called for element '+i+' ##');
console.info(Tag+'StrId for '+i+'is:'+AudioCapturerChangeInfoArray[i].streamId);
console.info(Tag+'CUid for '+i+'is:'+AudioCapturerChangeInfoArray[i].clientUid);
console.info(Tag+'Src for '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerInfo.source);
console.info(Tag+'Flag '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerInfo.capturerFlags);
console.info(Tag+'State for '+i+'is:'+AudioCapturerChangeInfoArray[i].capturerState);
var devDescriptor = AudioCapturerChangeInfoArray[i].deviceDescriptors;
for (let j=0;j<AudioCapturerChangeInfoArray[i].deviceDescriptors.length; j++) {
console.info(Tag+'Id:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].id);
console.info(Tag+'Type:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].deviceType);
console.info(Tag+'Role:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].deviceRole);
console.info(Tag+'Name:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].name);
console.info(Tag+'Addr:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].address);
console.info(Tag+'SR:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].sampleRates[0]);
console.info(Tag+'C'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].channelCounts[0]);
console.info(Tag+'CM:'+i+':'+AudioCapturerChangeInfoArray[i].deviceDescriptors[j].channelMasks);
}
if (AudioCapturerChangeInfoArray[i].capturerState == 1 && devDescriptor != null) {
resultFlag = true;
console.info(Tag+'[CAPTURER-CHANGE-ON-001] ResultFlag for element '+i +' is: '+ resultFlag);
}
}
});
```
...
...
@@ -1845,10 +2314,16 @@ audioManager.getAudioScene().then((value) => {
**系统能力:**
以下各项对应的系统能力均为SystemCapability.Multimedia.Audio.Device
| 名称 | 类型 | 可读 | 可写 | 说明 |
| ---------- | ------------------------- | ---- | ---- | ---------- |
| deviceRole |
[
DeviceRole
](
#devicerole
)
| 是 | 否 | 设备角色。 |
| deviceType |
[
DeviceType
](
#devicetype
)
| 是 | 否 | 设备类型。 |
| 名称 | 类型 | 可读 | 可写 | 说明 |
| -------------------------- | -------------------------- | ---- | ---- | ---------- |
| deviceRole |
[
DeviceRole
](
#devicerole
)
| 是 | 否 | 设备角色。 |
| deviceType |
[
DeviceType
](
#devicetype
)
| 是 | 否 | 设备类型。 |
| id
<sup>
9+
</sup>
| number | 是 | 否 | 设备id。 |
| name
<sup>
9+
</sup>
| string | 是 | 否 | 设备名称。 |
| address
<sup>
9+
</sup>
| string | 是 | 否 | 设备地址。 |
| sampleRates
<sup>
9+
</sup>
| Array
<
number
>
| 是 | 否 | 支持的采样率。 |
| channelCounts
<sup>
9+
</sup>
| Array
<
number
>
| 是 | 否 | 支持的通道数。 |
| channelMasks
<sup>
9+
</sup>
| Array
<
number
>
| 是 | 否 | 支持的通道掩码。 |
## AudioDeviceDescriptors
...
...
@@ -1860,23 +2335,22 @@ audioManager.getAudioScene().then((value) => {
import audio from '@ohos.multimedia.audio';
function displayDeviceProp(value) {
deviceRoleValue = value.deviceRole;
deviceTypeValue = value.deviceType;
deviceRoleValue = value.deviceRole;
deviceTypeValue = value.deviceType;
}
var deviceRoleValue = null;
var deviceTypeValue = null;
const promise = audio.getAudioManager().getDevices(1);
promise.then(function (value) {
console.info('AudioFrameworkTest: Promise: getDevices OUTPUT_DEVICES_FLAG');
value.forEach(displayDeviceProp);
if (deviceTypeValue != null && deviceRoleValue != null){
console.info('AudioFrameworkTest: Promise: getDevices : OUTPUT_DEVICES_FLAG : PASS');
}
else{
console.info('AudioFrameworkTest: Promise: getDevices : OUTPUT_DEVICES_FLAG : FAIL');
}
console.info('AudioFrameworkTest: Promise: getDevices OUTPUT_DEVICES_FLAG');
value.forEach(displayDeviceProp);
if (deviceTypeValue != null && deviceRoleValue != null){
console.info('AudioFrameworkTest: Promise: getDevices : OUTPUT_DEVICES_FLAG : PASS');
}
else{
console.info('AudioFrameworkTest: Promise: getDevices : OUTPUT_DEVICES_FLAG : FAIL');
}
});
```
...
...
@@ -1904,7 +2378,7 @@ getRendererInfo(callback: AsyncCallback<AudioRendererInfo\>): void
获取当前被创建的音频渲染器的信息,使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -1916,10 +2390,10 @@ getRendererInfo(callback: AsyncCallback<AudioRendererInfo\>): void
```
audioRenderer.getRendererInfo((err, rendererInfo) => {
console.log('Renderer GetRendererInfo:');
console.log('Renderer content:' + rendererInfo.content);
console.log('Renderer usage:' + rendererInfo.usage);
console.log('Renderer flags:' + rendererInfo.rendererFlags);
console.log('Renderer GetRendererInfo:');
console.log('Renderer content:' + rendererInfo.content);
console.log('Renderer usage:' + rendererInfo.usage);
console.log('Renderer flags:' + rendererInfo.rendererFlags);
});
```
...
...
@@ -1929,7 +2403,7 @@ getRendererInfo(): Promise<AudioRendererInfo\>
获取当前被创建的音频渲染器的信息,使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -1942,13 +2416,13 @@ getRendererInfo(): Promise<AudioRendererInfo\>
```
var resultFlag = true;
audioRenderer.getRendererInfo().then((rendererInfo) => {
console.log('Renderer GetRendererInfo:');
console.log('Renderer content:' + rendererInfo.content);
console.log('Renderer usage:' + rendererInfo.usage);
console.log('Renderer flags:' + rendererInfo.rendererFlags);
console.log('Renderer GetRendererInfo:');
console.log('Renderer content:' + rendererInfo.content);
console.log('Renderer usage:' + rendererInfo.usage);
console.log('Renderer flags:' + rendererInfo.rendererFlags);
}).catch((err) => {
console.log('AudioFrameworkRenderLog: RendererInfo :ERROR: '+err.message);
resultFlag = false;
console.log('AudioFrameworkRenderLog: RendererInfo :ERROR: '+err.message);
resultFlag = false;
});
```
...
...
@@ -1958,7 +2432,7 @@ getStreamInfo(callback: AsyncCallback<AudioStreamInfo\>): void
获取音频流信息,使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -1970,11 +2444,11 @@ getStreamInfo(callback: AsyncCallback<AudioStreamInfo\>): void
```
audioRenderer.getStreamInfo((err, streamInfo) => {
console.log('Renderer GetStreamInfo:');
console.log('Renderer sampling rate:' + streamInfo.samplingRate);
console.log('Renderer channel:' + streamInfo.channels);
console.log('Renderer format:' + streamInfo.sampleFormat);
console.log('Renderer encoding type:' + streamInfo.encodingType);
console.log('Renderer GetStreamInfo:');
console.log('Renderer sampling rate:' + streamInfo.samplingRate);
console.log('Renderer channel:' + streamInfo.channels);
console.log('Renderer format:' + streamInfo.sampleFormat);
console.log('Renderer encoding type:' + streamInfo.encodingType);
});
```
...
...
@@ -1984,7 +2458,7 @@ getStreamInfo(): Promise<AudioStreamInfo\>
获取音频流信息,使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -1996,13 +2470,13 @@ getStreamInfo(): Promise<AudioStreamInfo\>
```
audioRenderer.getStreamInfo().then((streamInfo) => {
console.log('Renderer GetStreamInfo:');
console.log('Renderer sampling rate:' + streamInfo.samplingRate);
console.log('Renderer channel:' + streamInfo.channels);
console.log('Renderer format:' + streamInfo.sampleFormat);
console.log('Renderer encoding type:' + streamInfo.encodingType);
console.log('Renderer GetStreamInfo:');
console.log('Renderer sampling rate:' + streamInfo.samplingRate);
console.log('Renderer channel:' + streamInfo.channels);
console.log('Renderer format:' + streamInfo.sampleFormat);
console.log('Renderer encoding type:' + streamInfo.encodingType);
}).catch((err) => {
console.log('ERROR: '+err.message);
console.log('ERROR: '+err.message);
});
```
...
...
@@ -2012,7 +2486,7 @@ start(callback: AsyncCallback<void\>): void
启动音频渲染器。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2024,11 +2498,11 @@ start(callback: AsyncCallback<void\>): void
```
audioRenderer.start((err) => {
if (err) {
console.error('Renderer start failed.');
} else {
console.info('Renderer start success.');
}
if (err) {
console.error('Renderer start failed.');
} else {
console.info('Renderer start success.');
}
});
```
...
...
@@ -2038,7 +2512,7 @@ start(): Promise<void\>
启动音频渲染器。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -2050,9 +2524,9 @@ start(): Promise<void\>
```
audioRenderer.start().then(() => {
console.log('Renderer started');
console.log('Renderer started');
}).catch((err) => {
console.log('ERROR: '+err.message);
console.log('ERROR: '+err.message);
});
```
...
...
@@ -2062,7 +2536,7 @@ pause(callback: AsyncCallback\<void>): void
暂停渲染。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2074,11 +2548,11 @@ pause(callback: AsyncCallback\<void>): void
```
audioRenderer.pause((err) => {
if (err) {
console.error('Renderer pause failed');
} else {
console.log('Renderer paused.');
}
if (err) {
console.error('Renderer pause failed');
} else {
console.log('Renderer paused.');
}
});
```
...
...
@@ -2088,7 +2562,7 @@ pause(): Promise\<void>
暂停渲染。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -2100,9 +2574,9 @@ pause(): Promise\<void>
```
audioRenderer.pause().then(() => {
console.log('Renderer paused');
console.log('Renderer paused');
}).catch((err) => {
console.log('ERROR: '+err.message);
console.log('ERROR: '+err.message);
});
```
...
...
@@ -2112,7 +2586,7 @@ drain(callback: AsyncCallback\<void>): void
检查缓冲区是否已被耗尽。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2124,11 +2598,11 @@ drain(callback: AsyncCallback\<void>): void
```
audioRenderer.drain((err) => {
if (err) {
console.error('Renderer drain failed');
} else {
console.log('Renderer drained.');
}
if (err) {
console.error('Renderer drain failed');
} else {
console.log('Renderer drained.');
}
});
```
...
...
@@ -2138,7 +2612,7 @@ drain(): Promise\<void>
检查缓冲区是否已被耗尽。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -2150,9 +2624,9 @@ drain(): Promise\<void>
```
audioRenderer.drain().then(() => {
console.log('Renderer drained successfully');
console.log('Renderer drained successfully');
}).catch((err) => {
console.log('ERROR: '+err.message);
console.log('ERROR: '+err.message);
});
```
...
...
@@ -2162,7 +2636,7 @@ stop(callback: AsyncCallback\<void>): void
停止渲染。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2174,11 +2648,11 @@ stop(callback: AsyncCallback\<void>): void
```
audioRenderer.stop((err) => {
if (err) {
console.error('Renderer stop failed');
} else {
console.log('Renderer stopped.');
}
if (err) {
console.error('Renderer stop failed');
} else {
console.log('Renderer stopped.');
}
});
```
...
...
@@ -2188,7 +2662,7 @@ stop(): Promise\<void>
停止渲染。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -2200,9 +2674,9 @@ stop(): Promise\<void>
```
audioRenderer.stop().then(() => {
console.log('Renderer stopped successfully');
console.log('Renderer stopped successfully');
}).catch((err) => {
console.log('ERROR: '+err.message);
console.log('ERROR: '+err.message);
});
```
...
...
@@ -2212,7 +2686,7 @@ release(callback: AsyncCallback\<void>): void
释放音频渲染器。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2224,11 +2698,11 @@ release(callback: AsyncCallback\<void>): void
```
audioRenderer.release((err) => {
if (err) {
console.error('Renderer release failed');
} else {
console.log('Renderer released.');
}
if (err) {
console.error('Renderer release failed');
} else {
console.log('Renderer released.');
}
});
```
...
...
@@ -2238,7 +2712,7 @@ release(): Promise\<void>
释放渲染器。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -2250,9 +2724,9 @@ release(): Promise\<void>
```
audioRenderer.release().then(() => {
console.log('Renderer released successfully');
console.log('Renderer released successfully');
}).catch((err) => {
console.log('ERROR: '+err.message);
console.log('ERROR: '+err.message);
});
```
...
...
@@ -2262,7 +2736,7 @@ write(buffer: ArrayBuffer, callback: AsyncCallback\<number>): void
写入缓冲区。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2279,36 +2753,35 @@ import fileio from '@ohos.fileio';
import featureAbility from '@ohos.ability.featureAbility'
var audioStreamInfo = {
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_48000,
channels: audio.AudioChannel.CHANNEL_2,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S32LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_48000,
channels: audio.AudioChannel.CHANNEL_2,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S32LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioRendererInfo = {
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.streamUsage.STREAM_USAGE_VOICE_COMMUNICATION
rendererFlags: 1
}
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.StreamUsage.STREAM_USAGE_VOICE_COMMUNICATION
rendererFlags: 1
var audioRendererOptions = {
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
}
var audioRenderer;
audio.createAudioRenderer(audioRendererOptions).then((data)=> {
audioRenderer = data;
console.info('AudioFrameworkRenderLog: AudioRenderer Created: SUCCESS');
}).catch((err) => {
console.info('AudioFrameworkRenderLog: AudioRenderer Created: ERROR: '+err.message);
});
audioRenderer = data;
console.info('AudioFrameworkRenderLog: AudioRenderer Created: SUCCESS');
}).catch((err) => {
console.info('AudioFrameworkRenderLog: AudioRenderer Created: ERROR: '+err.message);
});
var bufferSize;
audioRenderer.getBufferSize().then((data)=> {
console.info('AudioFrameworkRenderLog: getBufferSize: SUCCESS '+data);
bufferSize = data;
}).catch((err) => {
console.info.('AudioFrameworkRenderLog: getBufferSize: ERROR: '+err.message);
});
console.info('AudioFrameworkRenderLog: getBufferSize: SUCCESS '+data);
bufferSize = data;
}).catch((err) => {
console.info.('AudioFrameworkRenderLog: getBufferSize: ERROR: '+err.message);
});
console.info('Buffer size:'+bufferSize);
var context = featureAbility.getContext();
var path = await context.getCacheDir();
...
...
@@ -2317,11 +2790,11 @@ let ss = fileio.createStreamSync(filePath, 'r');
let buf = new ArrayBuffer(bufferSize);
ss.readSync(buf);
audioRenderer.write(buf, (err, writtenbytes) => {
if (writtenbytes < 0) {
console.error('write failed.');
} else {
console.log('Actual written bytes: ' + writtenbytes);
}
if (writtenbytes < 0) {
console.error('write failed.');
} else {
console.log('Actual written bytes: ' + writtenbytes);
}
});
```
...
...
@@ -2331,7 +2804,7 @@ write(buffer: ArrayBuffer): Promise\<number>
写入缓冲区。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -2347,36 +2820,36 @@ import fileio from '@ohos.fileio';
import featureAbility from '@ohos.ability.featureAbility'
var audioStreamInfo = {
samplingRate:audio.AudioSamplingRate.SAMPLE_RATE_48000,
channels:audio.AudioChannel.CHANNEL_2,
sampleFormat.
audio.AudioSampleFormat.SAMPLE_FORMAT_S32LE,
encodingType.
audio.AudioEncodingType.ENCODING_TYPE_RAW
samplingRate:audio.AudioSamplingRate.SAMPLE_RATE_48000,
channels:audio.AudioChannel.CHANNEL_2,
sampleFormat:
audio.AudioSampleFormat.SAMPLE_FORMAT_S32LE,
encodingType:
audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioRendererInfo = {
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.s
treamUsage.STREAM_USAGE_VOICE_COMMUNICATION,
rendererFlags: 1
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.S
treamUsage.STREAM_USAGE_VOICE_COMMUNICATION,
rendererFlags: 1
}
var audioRendererOptions = {
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
}
var audioRenderer;
audio.createAudioRenderer(audioRendererOptions).then((data) => {
audioRenderer = data;
console.info('AudioFrameworkRenderLog: AudioRenderer Created: SUCCESS');
}).catch((err) => {
console.info('AudioFrameworkRenderLog: AudioRenderer Created: ERROR: '+err.message);
});
audioRenderer = data;
console.info('AudioFrameworkRenderLog: AudioRenderer Created: SUCCESS');
}).catch((err) => {
console.info('AudioFrameworkRenderLog: AudioRenderer Created: ERROR: '+err.message);
});
var bufferSize;
audioRenderer.getBufferSize().then((data) => {
console.info('AudioFrameworkRenderLog: getBufferSize: SUCCESS '+data);
bufferSize = data;
}).catch((err) => {
console.info('AudioFrameworkRenderLog: getBufferSize: ERROR: '+err.message);
});
console.info('AudioFrameworkRenderLog: getBufferSize: SUCCESS '+data);
bufferSize = data;
}).catch((err) => {
console.info('AudioFrameworkRenderLog: getBufferSize: ERROR: '+err.message);
});
console.info('BufferSize: ' + bufferSize);
var context = featureAbility.getContext();
var path = await context.getCacheDir();
...
...
@@ -2385,11 +2858,11 @@ let ss = fileio.createStreamSync(filePath, 'r');
let buf = new ArrayBuffer(bufferSize);
ss.readSync(buf);
audioRenderer.write(buf).then((writtenbytes) => {
if (writtenbytes < 0) {
console.error('write failed.');
} else {
console.log('Actual written bytes: ' + writtenbytes);
}
if (writtenbytes < 0) {
console.error('write failed.');
} else {
console.log('Actual written bytes: ' + writtenbytes);
}
}).catch((err) => {
console.log('ERROR: '+err.message);
});
...
...
@@ -2401,7 +2874,7 @@ getAudioTime(callback: AsyncCallback\<number>): void
获取时间戳(从 1970 年 1 月 1 日开始)。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2413,7 +2886,7 @@ getAudioTime(callback: AsyncCallback\<number>): void
```
audioRenderer.getAudioTime((err, timestamp) => {
console.log('Current timestamp: ' + timestamp);
console.log('Current timestamp: ' + timestamp);
});
```
...
...
@@ -2423,7 +2896,7 @@ getAudioTime(): Promise\<number>
获取时间戳(从 1970 年 1 月 1 日开始)。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -2435,9 +2908,9 @@ getAudioTime(): Promise\<number>
```
audioRenderer.getAudioTime().then((timestamp) => {
console.log('Current timestamp: ' + timestamp);
console.log('Current timestamp: ' + timestamp);
}).catch((err) => {
console.log('ERROR: '+err.message);
console.log('ERROR: '+err.message);
});
```
...
...
@@ -2447,7 +2920,7 @@ getBufferSize(callback: AsyncCallback\<number>): void
获取音频渲染器的最小缓冲区大小。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2459,9 +2932,9 @@ getBufferSize(callback: AsyncCallback\<number>): void
```
var bufferSize = audioRenderer.getBufferSize(async(err, bufferSize) => {
if (err) {
console.error('getBufferSize error');
}
if (err) {
console.error('getBufferSize error');
}
});
```
...
...
@@ -2471,7 +2944,7 @@ getBufferSize(): Promise\<number>
获取音频渲染器的最小缓冲区大小。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -2486,35 +2959,35 @@ import audio from '@ohos.multimedia.audio';
import fileio from '@ohos.fileio';
var audioStreamInfo = {
samplingRate:audio.AudioSamplingRate.SAMPLE_RATE_48000,
channels:audio.AudioChannel.CHANNEL_2,
sampleFormat.
audio.AudioSampleFormat.SAMPLE_FORMAT_S32LE,
encodingType.
audio.AudioEncodingType.ENCODING_TYPE_RAW
samplingRate:audio.AudioSamplingRate.SAMPLE_RATE_48000,
channels:audio.AudioChannel.CHANNEL_2,
sampleFormat:
audio.AudioSampleFormat.SAMPLE_FORMAT_S32LE,
encodingType:
audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioRendererInfo = {
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.s
treamUsage.STREAM_USAGE_VOICE_COMMUNICATION,
rendererFlags: 1
content: audio.ContentType.CONTENT_TYPE_SPEECH,
usage: audio.S
treamUsage.STREAM_USAGE_VOICE_COMMUNICATION,
rendererFlags: 1
}
var audioRendererOptions = {
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
}
var audioRenderer;
audio.createAudioRenderer(audioRendererOptions).then((data) => {
audioRenderer = data;
console.info('AudioFrameworkRenderLog: AudioRenderer Created: SUCCESS');
}).catch((err) => {
console.info('AudioFrameworkRenderLog: AudioRenderer Created: ERROR: '+err.message);
});
audioRenderer = data;
console.info('AudioFrameworkRenderLog: AudioRenderer Created: SUCCESS');
}).catch((err) => {
console.info('AudioFrameworkRenderLog: AudioRenderer Created: ERROR: '+err.message);
});
var bufferSize;
audioRenderer.getBufferSize().then((data) => {
console.info('AudioFrameworkRenderLog: getBufferSize: SUCCESS '+data);
bufferSize=data;
console.info('AudioFrameworkRenderLog: getBufferSize: SUCCESS '+data);
bufferSize=data;
}).catch((err) => {
console.info('AudioFrameworkRenderLog: getBufferSize: ERROR: '+err.message);
console.info('AudioFrameworkRenderLog: getBufferSize: ERROR: '+err.message);
});
```
...
...
@@ -2524,7 +2997,7 @@ setRenderRate(rate: AudioRendererRate, callback: AsyncCallback\<void>): void
设置音频渲染速率。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2537,11 +3010,11 @@ setRenderRate(rate: AudioRendererRate, callback: AsyncCallback\<void>): void
```
audioRenderer.setRenderRate(audio.AudioRendererRate.RENDER_RATE_NORMAL, (err) => {
if (err) {
console.error('Failed to set params');
} else {
console.log('Callback invoked to indicate a successful render rate setting.');
}
if (err) {
console.error('Failed to set params');
} else {
console.log('Callback invoked to indicate a successful render rate setting.');
}
});
```
...
...
@@ -2551,7 +3024,7 @@ setRenderRate(rate: AudioRendererRate): Promise\<void>
设置音频渲染速率。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2569,9 +3042,9 @@ setRenderRate(rate: AudioRendererRate): Promise\<void>
```
audioRenderer.setRenderRate(audio.AudioRendererRate.RENDER_RATE_NORMAL).then(() => {
console.log('setRenderRate SUCCESS');
console.log('setRenderRate SUCCESS');
}).catch((err) => {
console.log('ERROR: '+err.message);
console.log('ERROR: '+err.message);
});
```
...
...
@@ -2581,7 +3054,7 @@ getRenderRate(callback: AsyncCallback\<AudioRendererRate>): void
获取当前渲染速率。使用callback方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2593,7 +3066,7 @@ getRenderRate(callback: AsyncCallback\<AudioRendererRate>): void
```
audioRenderer.getRenderRate((err, renderrate) => {
console.log('getRenderRate: ' + renderrate);
console.log('getRenderRate: ' + renderrate);
});
```
...
...
@@ -2603,7 +3076,7 @@ getRenderRate(): Promise\<AudioRendererRate>
获取当前渲染速率。使用Promise方式异步返回结果。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**返回值:**
...
...
@@ -2615,14 +3088,14 @@ getRenderRate(): Promise\<AudioRendererRate>
```
audioRenderer.getRenderRate().then((renderRate) => {
console.log('getRenderRate: ' + renderRate);
console.log('getRenderRate: ' + renderRate);
}).catch((err) => {
console.log('ERROR: '+err.message);
console.log('ERROR: '+err.message);
});
```
### setInterruptMode<sup>9+</sup>
setInterruptMode(
interruptM
ode: InterruptMode): Promise
<
void
>
setInterruptMode(
m
ode: InterruptMode): Promise
<
void
>
设置应用的焦点模型。使用Promise异步回调。
...
...
@@ -2630,9 +3103,9 @@ setInterruptMode(interruptMode: InterruptMode): Promise<void>
**参数:**
| 参数名 | 类型 | 必填
| 说明
|
| ---------- | ----------------------------------
- | ---- | ----------------------------------------------
---------- |
|
interruptMode |
[
InterruptMode
](
#InterruptMode
)
| 是 | 焦点模型。
|
| 参数名 | 类型 | 必填
| 说明
|
| ---------- | ----------------------------------
| ------ |
---------- |
|
mode |
[
InterruptMode
](
#InterruptMode
)
| 是 | 焦点模型。
|
**返回值:**
...
...
@@ -2643,14 +3116,32 @@ setInterruptMode(interruptMode: InterruptMode): Promise<void>
**示例:**
```
const audioManager = audio.getAudioManager();
audioManager.setInterruptMode(audio.InterruptMode.SHARE_MODE).then(() => {
console.log('Promise returned to indicate a successful volume setting.');
var audioStreamInfo = {
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_48000,
channels: audio.AudioChannel.CHANNEL_1,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioRendererInfo = {
content: audio.ContentType.CONTENT_TYPE_MUSIC,
usage: audio.StreamUsage.STREAM_USAGE_MEDIA,
rendererFlags: 0
}
var audioRendererOptions = {
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
}
let audioRenderer = await audio.createAudioRenderer(audioRendererOptions);
let mode = 0;
audioRenderer.setInterruptMode(mode).then(data=>{
console.log("setInterruptMode Success!");
}).catch(err=>{
console.log("setInterruptMode Fail:" + err.message);
});
```
### setInterruptMode<sup>9+</sup>
setInterruptMode(
interruptM
ode: InterruptMode, callback: Callback
\<
void>): void
setInterruptMode(
m
ode: InterruptMode, callback: Callback
\<
void>): void
设置应用的焦点模型。使用Callback回调返回执行结果。
...
...
@@ -2658,17 +3149,36 @@ setInterruptMode(interruptMode: InterruptMode, callback: Callback\<void>): void
**参数:**
| 参数名
| 类型 | 必填 | 说明
|
| -------
--- | ----------------------------------- | ---- | ------------------------------------------
-------------- |
|
interruptMode |
[
InterruptMode
](
#InterruptMode
)
| 是
| 焦点模型。|
|callback | Callback
\<
void>
| 是
|回调返回执行结果。|
| 参数名
| 类型 | 必填 | 说明
|
| -------
| ----------------------------------- | ------ |
-------------- |
|
mode |
[
InterruptMode
](
#InterruptMode
)
| 是
| 焦点模型。|
|callback | Callback
\<
void>
| 是
|回调返回执行结果。|
**示例:**
```
const audioManager = audio.getAudioManager();
audioManager.setInterruptMode(audio.InterruptMode.SHARE_MODE,()=>{
console.log('Callback returned to indicate a successful volume setting.');
var audioStreamInfo = {
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_48000,
channels: audio.AudioChannel.CHANNEL_1,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioRendererInfo = {
content: audio.ContentType.CONTENT_TYPE_MUSIC,
usage: audio.StreamUsage.STREAM_USAGE_MEDIA,
rendererFlags: 0
}
var audioRendererOptions = {
streamInfo: audioStreamInfo,
rendererInfo: audioRendererInfo
}
let audioRenderer = await audio.createAudioRenderer(audioRendererOptions);
let mode = 1;
audioRenderer.setInterruptMode(mode,(err,data)=>{
if(err){
console.log("setInterruptMode Fail:" + err.message);
}
console.log("setInterruptMode Success!");
});
```
### on('interrupt')<sup>9+</sup>
...
...
@@ -2677,7 +3187,7 @@ on(type: 'interrupt', callback: Callback\<InterruptEvent>): void
监听音频中断事件。使用callback获取中断事件。
**系统能力
**
:
SystemCapability.Multimedia.Audio.Renderer
**系统能力
:**
SystemCapability.Multimedia.Audio.Renderer
**参数:**
...
...
@@ -2692,49 +3202,48 @@ on(type: 'interrupt', callback: Callback\<InterruptEvent>): void
var isPlay;
var started;
audioRenderer.on('interrupt', async(interruptEvent) => {
if (interruptEvent.forceType == audio.InterruptForceType.INTERRUPT_FORCE) {
switch (interruptEvent.hintType) {
case audio.InterruptHint.INTERRUPT_HINT_PAUSE:
console.log('Force paused. Stop writing');
isPlay = false;
break;
case audio.InterruptHint.INTERRUPT_HINT_STOP:
console.log('Force stopped. Stop writing');
isPlay = false;
break;
}
} else if (interruptEvent.forceType == audio.InterruptForceType.INTERRUPT_SHARE) {
switch (interruptEvent.hintType) {
case audio.InterruptHint.INTERRUPT_HINT_RESUME:
console.log('Resume force paused renderer or ignore');
await audioRenderer.start().then(async function () {
console.info('AudioInterruptMusic: renderInstant started :SUCCESS ');
started = true;
}).catch((err) => {
console.info('AudioInterruptMusic: renderInstant start :ERROR : '+err.message);
started = false;
});
if (started) {
isPlay = true;
console.info('AudioInterruptMusic Renderer started : isPlay : '+isPlay);
} else {
console.error('AudioInterruptMusic Renderer start failed');
}
break;
case audio.InterruptHint.INTERRUPT_HINT_PAUSE:
console.log('Choose to pause or ignore');
if (isPlay == true) {
isPlay == false;
console.info('AudioInterruptMusic: Media PAUSE : TRUE');
}
else {
isPlay = true;
console.info('AudioInterruptMusic: Media PLAY : TRUE');
}
break;
if (interruptEvent.forceType == audio.InterruptForceType.INTERRUPT_FORCE) {
switch (interruptEvent.hintType) {
case audio.InterruptHint.INTERRUPT_HINT_PAUSE:
console.log('Force paused. Stop writing');
isPlay = false;
break;
case audio.InterruptHint.INTERRUPT_HINT_STOP:
console.log('Force stopped. Stop writing');
isPlay = false;
break;
}
} else if (interruptEvent.forceType == audio.InterruptForceType.INTERRUPT_SHARE) {
switch (interruptEvent.hintType) {
case audio.InterruptHint.INTERRUPT_HINT_RESUME:
console.log('Resume force paused renderer or ignore');
await audioRenderer.start().then(async function () {
console.info('AudioInterruptMusic: renderInstant started :SUCCESS ');
started = true;
}).catch((err) => {
console.info('AudioInterruptMusic: renderInstant start :ERROR : '+err.message);
started = false;
});
if (started) {
isPlay = true;
console.info('AudioInterruptMusic Renderer started : isPlay : '+isPlay);
} else {
console.error('AudioInterruptMusic Renderer start failed');
}
break;
case audio.InterruptHint.INTERRUPT_HINT_PAUSE:
console.log('Choose to pause or ignore');
if (isPlay == true) {
isPlay == false;
console.info('AudioInterruptMusic: Media PAUSE : TRUE');
}else {
isPlay = true;
console.info('AudioInterruptMusic: Media PLAY : TRUE');
}
break;
}
});
}
});
```
### on('markReach')<sup>8+</sup>
...
...
@@ -2757,9 +3266,9 @@ on(type: 'markReach', frame: number, callback: (position: number) => {}): void
```
audioRenderer.on('markReach', 1000, (position) => {
if (position == 1000) {
console.log('ON Triggered successfully');
}
if (position == 1000) {
console.log('ON Triggered successfully');
}
});
```
...
...
@@ -2804,9 +3313,9 @@ on(type: "periodReach", frame: number, callback: (position: number) => {}): void
```
audioRenderer.on('periodReach', 1000, (position) => {
if (position == 1000) {
console.log('ON Triggered successfully');
}
if (position == 1000) {
console.log('ON Triggered successfully');
}
});
```
...
...
@@ -2849,12 +3358,12 @@ on(type: 'stateChange', callback: Callback<AudioState\>): void
```
audioRenderer.on('stateChange', (state) => {
if (state == 1) {
console.log("audio renderer state is: STATE_PREPARED");
}
if (state == 2) {
console.log("audio renderer state is: STATE_RUNNING");
}
if (state == 1) {
console.log("audio renderer state is: STATE_PREPARED");
}
if (state == 2) {
console.log("audio renderer state is: STATE_RUNNING");
}
});
```
...
...
@@ -2894,13 +3403,13 @@ getCapturerInfo(callback: AsyncCallback<AudioCapturerInfo\>): void
```
audioCapturer.getCapturerInfo((err, capturerInfo) => {
if (err) {
console.error('Failed to get capture info');
} else {
console.log('Capturer getCapturerInfo:');
console.log('Capturer source:' + capturerInfo.source);
console.log('Capturer flags:' + capturerInfo.capturerFlags);
}
if (err) {
console.error('Failed to get capture info');
} else {
console.log('Capturer getCapturerInfo:');
console.log('Capturer source:' + capturerInfo.source);
console.log('Capturer flags:' + capturerInfo.capturerFlags);
}
});
```
...
...
@@ -2923,16 +3432,16 @@ getCapturerInfo(): Promise<AudioCapturerInfo\>
```
audioCapturer.getCapturerInfo().then((audioParamsGet) => {
if (audioParamsGet != undefined) {
console.info('AudioFrameworkRecLog: Capturer CapturerInfo:');
console.info('AudioFrameworkRecLog: Capturer SourceType:' + audioParamsGet.source);
console.info('AudioFrameworkRecLog: Capturer capturerFlags:' + audioParamsGet.capturerFlags);
}else {
console.info('AudioFrameworkRecLog: audioParamsGet is : '+audioParamsGet);
console.info('AudioFrameworkRecLog: audioParams getCapturerInfo are incorrect: ');
}
if (audioParamsGet != undefined) {
console.info('AudioFrameworkRecLog: Capturer CapturerInfo:');
console.info('AudioFrameworkRecLog: Capturer SourceType:' + audioParamsGet.source);
console.info('AudioFrameworkRecLog: Capturer capturerFlags:' + audioParamsGet.capturerFlags);
}else {
console.info('AudioFrameworkRecLog: audioParamsGet is : '+audioParamsGet);
console.info('AudioFrameworkRecLog: audioParams getCapturerInfo are incorrect: ');
}
}).catch((err) => {
console.log('AudioFrameworkRecLog: CapturerInfo :ERROR: '+err.message);
console.log('AudioFrameworkRecLog: CapturerInfo :ERROR: '+err.message);
});
```
...
...
@@ -2954,15 +3463,15 @@ getStreamInfo(callback: AsyncCallback<AudioStreamInfo\>): void
```
audioCapturer.getStreamInfo((err, streamInfo) => {
if (err) {
console.error('Failed to get stream info');
} else {
console.log('Capturer GetStreamInfo:');
console.log('Capturer sampling rate:' + streamInfo.samplingRate);
console.log('Capturer channel:' + streamInfo.channels);
console.log('Capturer format:' + streamInfo.sampleFormat);
console.log('Capturer encoding type:' + streamInfo.encodingType);
}
if (err) {
console.error('Failed to get stream info');
} else {
console.log('Capturer GetStreamInfo:');
console.log('Capturer sampling rate:' + streamInfo.samplingRate);
console.log('Capturer channel:' + streamInfo.channels);
console.log('Capturer format:' + streamInfo.sampleFormat);
console.log('Capturer encoding type:' + streamInfo.encodingType);
}
});
```
...
...
@@ -2984,13 +3493,13 @@ getStreamInfo(): Promise<AudioStreamInfo\>
```
audioCapturer.getStreamInfo().then((audioParamsGet) => {
console.info('getStreamInfo:');
console.info('sampleFormat:' + audioParamsGet.sampleFormat);
console.info('samplingRate:' + audioParamsGet.samplingRate);
console.info('channels:' + audioParamsGet.channels);
console.info('encodingType:' + audioParamsGet.encodingType);
console.info('getStreamInfo:');
console.info('sampleFormat:' + audioParamsGet.sampleFormat);
console.info('samplingRate:' + audioParamsGet.samplingRate);
console.info('channels:' + audioParamsGet.channels);
console.info('encodingType:' + audioParamsGet.encodingType);
}).catch((err) => {
console.log('getStreamInfo :ERROR: ' + err.message);
console.log('getStreamInfo :ERROR: ' + err.message);
});
```
...
...
@@ -3002,7 +3511,7 @@ start(callback: AsyncCallback<void\>): void
**系统能力:**
SystemCapability.Multimedia.Audio.Capturer
**参数**
**参数
:
**
| 参数名 | 类型 | 必填 | 说明 |
| :------- | :------------------- | :--- | :----------------------------- |
...
...
@@ -3012,11 +3521,11 @@ start(callback: AsyncCallback<void\>): void
```
audioCapturer.start((err) => {
if (err) {
console.error('Capturer start failed.');
} else {
console.info('Capturer start success.');
}
if (err) {
console.error('Capturer start failed.');
} else {
console.info('Capturer start success.');
}
});
```
...
...
@@ -3042,35 +3551,36 @@ import audio from '@ohos.multimedia.audio';
import fileio from '@ohos.fileio';
var audioStreamInfo = {
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_2,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
samplingRate: audio.AudioSamplingRate.SAMPLE_RATE_44100,
channels: audio.AudioChannel.CHANNEL_2,
sampleFormat: audio.AudioSampleFormat.SAMPLE_FORMAT_S16LE,
encodingType: audio.AudioEncodingType.ENCODING_TYPE_RAW
}
var audioCapturerInfo = {
source: audio.SourceType.SOURCE_TYPE_MIC,
capturerFlags =
1
source: audio.SourceType.SOURCE_TYPE_MIC,
capturerFlags:
1
}
var audioCapturer;
var stateFlag;
audio.createAudioCapturer(audioCapturerOptions).then((data) => {
audioCapturer = data;
console.info('AudioFrameworkRecLog: AudioCapturer Created: SUCCESS');
}).catch((err) => {
console.info('AudioFrameworkRecLog: AudioCapturer Created: ERROR: '+err.message);
});
audioCapturer = data;
console.info('AudioFrameworkRecLog: AudioCapturer Created: SUCCESS');
}).catch((err) => {
console.info('AudioFrameworkRecLog: AudioCapturer Created: ERROR: '+err.message);
});
audioCapturer.start().then(() => {
console.info('AudioFrameworkRecLog: ---------START---------');
console.info('AudioFrameworkRecLog: Capturer started: SUCCESS');
console.info('AudioFrameworkRecLog: AudioCapturer: STATE: '+audioCapturer.state);
console.info('AudioFrameworkRecLog: Capturer started: SUCCESS ');
if ((audioCapturer.state == audio.AudioState.STATE_RUNNING)) {
console.info('AudioFrameworkRecLog: AudioCapturer is in Running State');
}
console.info('AudioFrameworkRecLog: ---------START---------');
console.info('AudioFrameworkRecLog: Capturer started: SUCCESS');
console.info('AudioFrameworkRecLog: AudioCapturer: STATE: '+audioCapturer.state);
console.info('AudioFrameworkRecLog: Capturer started: SUCCESS ');
if ((audioCapturer.state == audio.AudioState.STATE_RUNNING)) {
console.info('AudioFrameworkRecLog: AudioCapturer is in Running State');
}
}).catch((err) => {
console.info('AudioFrameworkRecLog: Capturer start :ERROR : '+err.message);
stateFlag=false;
console.info('AudioFrameworkRecLog: Capturer start :ERROR : '+err.message);
stateFlag=false;
});
```
...
...
@@ -3092,11 +3602,11 @@ stop(callback: AsyncCallback<void\>): void
```
audioCapturer.stop((err) => {
if (err) {
console.error('Capturer stop failed');
} else {
console.log('Capturer stopped.');
}
if (err) {
console.error('Capturer stop failed');
} else {
console.log('Capturer stopped.');
}
});
```
...
...
@@ -3119,13 +3629,13 @@ stop(): Promise<void\>
```
audioCapturer.stop().then(() => {
console.info('AudioFrameworkRecLog: ---------STOP RECORD---------');
console.info('AudioFrameworkRecLog: Capturer stopped: SUCCESS');
if ((audioCapturer.state == audio.AudioState.STATE_STOPPED)){
console.info('AudioFrameworkRecLog: State is Stopped': ');
}
console.info('AudioFrameworkRecLog: ---------STOP RECORD---------');
console.info('AudioFrameworkRecLog: Capturer stopped: SUCCESS');
if ((audioCapturer.state == audio.AudioState.STATE_STOPPED)){
console.info('AudioFrameworkRecLog: State is Stopped': ');
}
}).catch((err) => {
console.info('AudioFrameworkRecLog: Capturer stop: ERROR: '+err.message);
console.info('AudioFrameworkRecLog: Capturer stop: ERROR: '+err.message);
});
```
...
...
@@ -3147,11 +3657,11 @@ release(callback: AsyncCallback<void\>): void
```
audioCapturer.release((err) => {
if (err) {
console.error('capturer release failed');
} else {
console.log('capturer released.');
}
if (err) {
console.error('capturer release failed');
} else {
console.log('capturer released.');
}
});
```
...
...
@@ -3173,13 +3683,14 @@ release(): Promise<void\>
**示例:**
```
var stateFlag;
audioCapturer.release().then(() => {
console.info('AudioFrameworkRecLog: ---------RELEASE RECORD---------');
console.info('AudioFrameworkRecLog: Capturer release : SUCCESS');
console.info('AudioFrameworkRecLog: AudioCapturer : STATE : '+audioCapturer.state);
console.info('AudioFrameworkRecLog: stateFlag : '+stateFlag);
console.info('AudioFrameworkRecLog: ---------RELEASE RECORD---------');
console.info('AudioFrameworkRecLog: Capturer release : SUCCESS');
console.info('AudioFrameworkRecLog: AudioCapturer : STATE : '+audioCapturer.state);
console.info('AudioFrameworkRecLog: stateFlag : '+stateFlag);
}).catch((err) => {
console.info('AudioFrameworkRecLog: Capturer stop: ERROR: '+err.message);
console.info('AudioFrameworkRecLog: Capturer stop: ERROR: '+err.message);
});
```
...
...
@@ -3192,7 +3703,7 @@ read(size: number, isBlockingRead: boolean, callback: AsyncCallback<ArrayBuffer\
**系统能力:**
SystemCapability.Multimedia.Audio.Capturer
**参数**
**参数
:
**
| 参数名 | 类型 | 必填 | 说明 |
| :------------- | :-------------------------- | :--- | :------------------------------- |
...
...
@@ -3205,15 +3716,15 @@ read(size: number, isBlockingRead: boolean, callback: AsyncCallback<ArrayBuffer\
```
var bufferSize;
audioCapturer.getBufferSize().then((data) => {
console.info('AudioFrameworkRecLog: getBufferSize: SUCCESS '+data);
bufferSize = data;
}).catch((err) => {
console.info('AudioFrameworkRecLog: getBufferSize: EROOR: '+err.message);
});
console.info('AudioFrameworkRecLog: getBufferSize: SUCCESS '+data);
bufferSize = data;
}).catch((err) => {
console.info('AudioFrameworkRecLog: getBufferSize: EROOR: '+err.message);
});
audioCapturer.read(bufferSize, true, async(err, buffer) => {
if (!err) {
console.log("Success in reading the buffer data");
}
if (!err) {
console.log("Success in reading the buffer data");
}
});
```
...
...
@@ -3244,16 +3755,16 @@ read(size: number, isBlockingRead: boolean): Promise<ArrayBuffer\>
```
var bufferSize;
audioCapturer.getBufferSize().then((data) => {
console.info('AudioFrameworkRecLog: getBufferSize: SUCCESS '+data);
bufferSize = data;
}).catch((err) => {
console.info('AudioFrameworkRecLog: getBufferSize: ERROR '+err.message);
});
console.info('AudioFrameworkRecLog: getBufferSize: SUCCESS '+data);
bufferSize = data;
}).catch((err) => {
console.info('AudioFrameworkRecLog: getBufferSize: ERROR '+err.message);
});
console.info('Buffer size: ' + bufferSize);
audioCapturer.read(bufferSize, true).then((buffer) => {
console.info('buffer read successfully');
console.info('buffer read successfully');
}).catch((err) => {
console.info('ERROR : '+err.message);
console.info('ERROR : '+err.message);
});
```
...
...
@@ -3276,7 +3787,7 @@ getAudioTime(callback: AsyncCallback<number\>): void
```
audioCapturer.getAudioTime((err, timestamp) => {
console.log('Current timestamp: ' + timestamp);
console.log('Current timestamp: ' + timestamp);
});
```
...
...
@@ -3299,9 +3810,9 @@ getAudioTime(): Promise<number\>
```
audioCapturer.getAudioTime().then((audioTime) => {
console.info('AudioFrameworkRecLog: AudioCapturer getAudioTime : Success' + audioTime );
console.info('AudioFrameworkRecLog: AudioCapturer getAudioTime : Success' + audioTime );
}).catch((err) => {
console.info('AudioFrameworkRecLog: AudioCapturer Created : ERROR : '+err.message);
console.info('AudioFrameworkRecLog: AudioCapturer Created : ERROR : '+err.message);
});
```
...
...
@@ -3324,14 +3835,14 @@ getBufferSize(callback: AsyncCallback<number\>): void
```
audioCapturer.getBufferSize((err, bufferSize) => {
if (!err) {
console.log('BufferSize : ' + bufferSize);
audioCapturer.read(bufferSize, true).then((buffer) => {
console.info('Buffer read is ' + buffer );
}).catch((err) => {
console.info('AudioFrameworkRecLog: AudioCapturer Created : ERROR : '+err.message);
});
}
if (!err) {
console.log('BufferSize : ' + bufferSize);
audioCapturer.read(bufferSize, true).then((buffer) => {
console.info('Buffer read is ' + buffer );
}).catch((err) => {
console.info('AudioFrameworkRecLog: AudioCapturer Created : ERROR : '+err.message);
});
}
});
```
...
...
@@ -3355,10 +3866,10 @@ getBufferSize(): Promise<number\>
```
var bufferSize;
audioCapturer.getBufferSize().then((data) => {
console.info('AudioFrameworkRecLog: getBufferSize :SUCCESS '+ data);
bufferSize = data;
console.info('AudioFrameworkRecLog: getBufferSize :SUCCESS '+ data);
bufferSize = data;
}).catch((err) => {
console.info('AudioFrameworkRecLog: getBufferSize :ERROR : '+ err.message);
console.info('AudioFrameworkRecLog: getBufferSize :ERROR : '+ err.message);
});
```
...
...
@@ -3383,9 +3894,9 @@ on(type: 'markReach', frame: number, callback: (position: number) => {}): void
```
audioCapturer.on('markReach', 1000, (position) => {
if (position == 1000) {
console.log('ON Triggered successfully');
}
if (position == 1000) {
console.log('ON Triggered successfully');
}
});
```
...
...
@@ -3429,9 +3940,9 @@ on(type: "periodReach", frame: number, callback: (position: number) => {}): void
```
audioCapturer.on('periodReach', 1000, (position) => {
if (position == 1000) {
console.log('ON Triggered successfully');
}
if (position == 1000) {
console.log('ON Triggered successfully');
}
});
```
...
...
@@ -3474,11 +3985,11 @@ on(type: 'stateChange', callback: Callback<AudioState\>): void
```
audioCapturer.on('stateChange', (state) => {
if (state == 1) {
console.log("audio capturer state is: STATE_PREPARED");
}
if (state == 2) {
console.log("audio capturer state is: STATE_RUNNING");
}
if (state == 1) {
console.log("audio capturer state is: STATE_PREPARED");
}
if (state == 2) {
console.log("audio capturer state is: STATE_RUNNING");
}
});
```
\ No newline at end of file
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录