anyLive 是 anyRTC 开源的推拉流项目。采用跨平台架构设计,一套代码支持Android、iOS、Windows、Mac、Ubuntu等平台。本文主要介绍anyLive iOS平台的实现。
源码下载- 源码下载(https://github.com/anyrtcIO-Community/anyRTC-RTMP-OpenSource)
开发环境
- 开发工具:Xcode13 真机运行
- 开发语言:Objective-C、Swift
- 实现:推拉流。
系统 |
编译环境 |
CPU架构 |
Android 4.4及以上 |
Android Studio、NDK |
armeabi-v7a、arm64-v8a |
iOS 9.0及以上 |
Xcode13 |
arm64 |
Windows 7及以上 |
VS2015,VS2017 |
x86、x86-64 |
anyLive 实现了推流、拉流、屏幕共享、美颜等功能。
示例代码效果展示
代码实现
var menus = [
[MenuItem(imageName: "icon_push", title: "直播推流", subTitle: "采用WebRTC底层架构,支持RTMP/HLS/HTTP-FLV")],
[MenuItem(imageName: "icon_pull", title: "直播拉流(播放)", subTitle: "低功直播播放器,支持软硬解切换,横竖切换、低延迟等")],
[MenuItem(imageName: "icon_video", title: "小视频播放", subTitle: "支持首屏秒开、清晰度无缝切换、码率自适应等多种特性")]
]
let identifier = "ARLiveMainCell"
lazy var identifierArr: [String] = {
["Live_JoinVC", "Player_JoinVC", "Video_JoinVC"]
}()
override func viewDidLoad() {
super.viewDidLoad()
// Uncomment the following line to preserve selection between presentations
// self.clearsSelectionOnViewWillAppear = false
// Uncomment the following line to display an Edit button in the navigation bar for this view controller.
// self.navigationItem.rightBarButtonItem = self.editButtonItem
let label = UILabel(frame: CGRectZero)
label.textColor = UIColor(hexString: "#C4C4CE")
label.font = UIFont(name: PingFang, size: 12)
label.textAlignment = .center
label.text = "Power by anyRTC"
view.addSubview(label)
liveEngine = ARLiveEngineKit(delegate: nil)
}
override func viewWillAppear(_ animated: Bool) {
super.viewWillAppear(animated)
navigationController?.setNavigationBarHidden(true, animated: true)
}
// MARK: - Table view data source
override func numberOfSections(in tableView: UITableView) -> Int {
return menus.count
}
override func tableView(_ tableView: UITableView, numberOfRowsInSection section: Int) -> Int {
// #warning Incomplete implementation, return the number of rows
return menus[section].count
}
override func tableView(_ tableView: UITableView, cellForRowAt indexPath: IndexPath) -> UITableViewCell {
let cell: ARMainCell = tableView.dequeueReusableCell(withIdentifier: identifier, for: indexPath) as! ARMainCell
// Configure the cell...
let menuItem = menus[indexPath.section][indexPath.row]
cell.mainImageView.image = UIImage(named: menuItem.imageName)
cell.mainLabel.text = menuItem.title
cell.subLabel.text = menuItem.subTitle
cell.expectedImageView.isHidden = (indexPath.section != 2)
return cell
}
override func tableView(_ tableView: UITableView, didSelectRowAt indexPath: IndexPath) {
if indexPath.section != 2 {
guard let vc = storyboard?.instantiateViewController(withIdentifier: identifierArr[indexPath.section]) else { return }
navigationController?.pushViewController(vc, animated: true)
} else {
ARToast.showText(text: " Please look forward!", duration: 1.0)
}
代码实现
func initializePusher() {
/// 实例化推流对象
livePusher = liveEngine!.createArLivePusher()
livePusher.setDelegate(self)
/// 设置推流视频编码参数
let param = ARLiveVideoEncoderParam(resolution!)
livePusher.setVideoQuality(param)
livePusher.startCamera(true)
livePusher.startMicrophone()
/// 设置本地摄像头预览 View
livePusher.setupCameraRender(renderView)
livePusher.setRenderFill(.fill)
/// 开始推流
livePusher.startPush(pushUrl)
}
// MARK: - ARLivePushDelegate
extension ArLiveViewController: ARLivePushDelegate {
func onError(_ code: ARLiveCode, message msg: String?, extraInfo: [AnyHashable: Any]?) {
/// 直播推流器错误通知,推流器出现错误时,会回调该通知
Logger.log(message: "onError \(code.rawValue)", level: .error)
}
func onWarning(_ code: ARLiveCode, message msg: String?, extraInfo: [AnyHashable: Any]?) {
/// 直播推流器警告通知
Logger.log(message: "onWarning \(code.rawValue)", level: .warning)
}
func onCaptureFirstAudioFrame() {
/// 首帧音频采集完成的回调通知
Logger.log(message: "onCaptureFirstAudioFrame", level: .info)
}
func onCaptureFirstVideoFrame() {
/// 首帧视频采集完成的回调通知
Logger.log(message: "onCaptureFirstVideoFrame", level: .info)
}
func onMicrophoneVolumeUpdate(_ volume: Int) {
/// 麦克风采集音量值回调
Logger.log(message: "onMicrophoneVolumeUpdate volume = \(volume)", level: .info)
}
func onPushStatusUpdate(_ status: ARLivePushStatus, message msg: String?, extraInfo: [AnyHashable: Any]?) {
/// 推流器连接状态回调通知
Logger.log(message: "onPushStatusUpdate status = \(status.rawValue)", level: .info)
stateLabel.text = "\(status.description)"
}
func onStatisticsUpdate(_ statistics: ARLivePusherStatistics) {
/// 直播推流器统计数据回调
// Logger.log(message: "onStatisticsUpdate width = \(statistics.width), height = \(statistics.height), fps = \(statistics.fps), videoBitrate = \(statistics.videoBitrate), audioBitrate = \(statistics.audioBitrate)", level: .info)
}
func onSnapshotComplete(_ image: UIImage) {
/// 截图回调
Logger.log(message: "onSnapshotComplete", level: .info)
}
}
代码实现
func initializePlayer() {
/// 创建拉流实例对象
livePlayer = liveEngine!.createArLivePlayer()
livePlayer.setDelegate(self)
/// 设置播放器的视频渲染 View
livePlayer.setRenderView(renderView)
livePlayer.setRenderFill(renderMode)
/// 设置播放器缓存自动调整的最小和最大时间 ( 单位:秒 )
livePlayer.setCacheParams(1.0, maxTime: 100)
/// 开始播放音视频流
livePlayer.startPlay(pullUrl)
}
// MARK: - ARLivePlayDelegate
extension ArPlayerViewController: ARLivePlayDelegate {
func onError(_ player: ARLivePlayer, code: ARLiveCode, message msg: String?, extraInfo: [AnyHashable: Any]?) {
/// 直播播放器错误通知,播放器出现错误时,会回调该通知
Logger.log(message: "onError code = \(code.rawValue)", level: .info)
}
func onWarning(_ player: ARLivePlayer, code: ARLiveCode, message msg: String?, extraInfo: [AnyHashable: Any]?) {
/// 直播播放器警告通知
Logger.log(message: "onWarning code = \(code.rawValue)", level: .info)
}
func onVideoPlayStatusUpdate(_ player: ARLivePlayer, status: ARLivePlayStatus, reason: ARLiveStatusChangeReason, extraInfo: [AnyHashable: Any]?) {
/// 直播播放器视频状态变化通知
Logger.log(message: "onVideoPlayStatusUpdate status = \(status.rawValue), reason = \(reason.rawValue)", level: .info)
liveStatus = status
stateLabel.text = "\(status.description)"
}
func onAudioPlayStatusUpdate(_ player: ARLivePlayer, status: ARLivePlayStatus, reason: ARLiveStatusChangeReason, extraInfo: [AnyHashable: Any]?) {
/// 直播播放器音频状态变化通知
Logger.log(message: "onAudioPlayStatusUpdate status = \(status.rawValue) reason = \(reason.rawValue)", level: .info)
}
func onPlayoutVolumeUpdate(_ player: ARLivePlayer, volume: Int) {
/// 播放器音量大小回调
Logger.log(message: "onPlayoutVolumeUpdate volume = \(volume)", level: .info)
}
func onStatisticsUpdate(_ player: ARLivePlayer, statistics: ARLivePlayerStatistics?) {
/// 直播播放器统计数据回调
if statistics != nil {
Logger.log(message: "onStatisticsUpdate width = \(statistics!.width), height =\(statistics!.height), fps = \(statistics!.fps), videoBitrate = \(statistics!.videoBitrate), audioBitrate = \(statistics!.audioBitrate)", level: .info)
}
}
func onSnapshotComplete(_ player: ARLivePlayer, image: UIImage) {
/// 截图回调
UIImageWriteToSavedPhotosAlbum(image, self, #selector(saveImage(image:didFinishSavingWithError:contextInfo:)), nil)
NSObject.cancelPreviousPerformRequests(withTarget: self, selector: #selector(removeSnapshot), object: nil)
snapImageView.image = image
let imageWidth = image.size.width/2
let imageHeight = image.size.height/2
snapImageView.frame = CGRect(x: ARScreenWidth - imageWidth - 24, y: 150, width: imageWidth, height: imageHeight)
view.addSubview(snapImageView)
perform(#selector(removeSnapshot), with: nil, afterDelay: 2)
Logger.log(message: "onSnapshotComplete sucess, imageWidth = \(image.size.width), imageHeight = \(image.size.height)", level: .info)
}
func onRenderVideoFrame(_ player: ARLivePlayer, frame videoFrame: ARLiveVideoFrame?) {
/// 自定义视频渲染回调
Logger.log(message: "onRenderVideoFrame", level: .info)
}
func onReceiveSeiMessage(_ player: ARLivePlayer, payloadType: Int32, data: Data?) {
/// 收到 SEI 消息的回调
Logger.log(message: "onReceiveSeiMessage payloadType = \(payloadType)", level: .info)
}
}
代码实现
override func processSampleBuffer(_ sampleBuffer: CMSampleBuffer, with sampleBufferType: RPSampleBufferType) {
DispatchQueue.main.async {
switch sampleBufferType {
case RPSampleBufferType.video:
// Handle video sample buffer
ARUploader.sendVideoBuffer(sampleBuffer)
case RPSampleBufferType.audioApp:
// Handle audio sample buffer for app audio
ARUploader.sendAudioAppBuffer(sampleBuffer)
case RPSampleBufferType.audioMic:
// Handle audio sample buffer for mic audio
ARUploader.sendAudioMicBuffer(sampleBuffer)
break
@unknown default:
// Handle other sample buffer types
fatalError("Unknown type of sample buffer")
}
}
}
private static let liverPusher: ARLivePusher = {
let livePusher = liveEngine.createArLivePusher()
let screenSize = UIScreen.main.currentMode?.size
let screenWidth = screenSize?.width
let screenHeight = screenSize?.height
/// 设置推流视频编码参数
let videoParam = ARLiveVideoEncoderParam()
videoParam.videoResolution = .resolution640x480
videoParam.videoResolutionMode = .portrait
videoParam.videoScaleMode = .fit
livePusher.setVideoQuality(videoParam)
livePusher.startMicrophone()
/// 开启自采集
livePusher.enableCustomAudioCapture(true)
livePusher.enableCustomVideoCapture(true)
/// 开始推流
livePusher.startPush(<#T##String#>)
return livePusher
}()
static func sendAudioAppBuffer(_ sampleBuffer: CMSampleBuffer) {
ARAudioTube.liverPusher(liverPusher, pushAudioCMSampleBuffer: sampleBuffer, resampleRate: audioSampleRate, type: .app)
}
static func sendAudioMicBuffer(_ sampleBuffer: CMSampleBuffer) {
ARAudioTube.liverPusher(liverPusher, pushAudioCMSampleBuffer: sampleBuffer, resampleRate: audioSampleRate, type: .mic)
}
最后,因时间有限项目中还存在一些bug和待完善的功能点。仅供参考,欢迎大家fork。有不足之处欢迎大家指出issues。最后再贴一下 Github开源下载地址。
Github开源下载地址(https://github.com/anyrtcIO-Community/anyRTC-RTMP-OpenSource)。
如果觉得不错,希望点个star~
,