Amazon Chime SDK Project Board
Note: If building with the SDK source code, the
development
branch contains bleeding-edge changes that may not build with the publically available Chime media library or may not be as stable as public releases.
The Amazon Chime SDK for iOS makes it easy to add collaborative audio calling, video calling, and screen share viewing features to iOS applications by using the same infrastructure services that power meetings on the Amazon Chime service.
This Amazon Chime SDK for iOS works by connecting to meeting session resources that you have created in your AWS account. The SDK has everything you need to build custom calling and collaboration experiences in your iOS application, including methods to: configure meeting sessions, list and select audio devices, switch video devices, start and stop screen share viewing, receive callbacks when media events occur such as volume changes, and manage meeting features such as audio mute and video tile bindings.
We also have an Amazon Chime SDK Project Board where you can find community requests and their statuses.
To get started, see the following resources:
And review the following guides:
- API Overview
- Getting Started
- Frequently Asked Questions (FAQ)
- Custom Video Sources, Processors, and Sinks
- Video Pagination with Active Speaker-Based Policy
- Content Share
- Meeting Events
- Event Ingestion
- Configuring Remote Video Subscription
- Background Video Filters
You can integrate Amazon Chime SDK in your project from SPM, CocoaPods or binaries through Github release.
For the purpose of setup, your project's root folder (where you can find your .xcodeproj
file) will be referred to as root
.
-
The Amazon Chime SDK is available through CocoaPods. If you have not installed CocoaPods, install CocoaPods by running the command:
$ gem install cocoapods $ pod setup
Depending on your system settings, you may have to use sudo for installing cocoapods as follows:
$ sudo gem install cocoapods $ pod setup
-
In root directory (the directory where your *.xcodeproj file is), run the following to create a Podfile in your project:
$ pod init
-
Edit the
Podfile
to includeAmazonChimeSDK
into your project:target 'YourTarget' do pod 'AmazonChimeSDK' ... end
If you don't need video and content share functionality, or software video codec support, you can use
AmazonChimeSDKNoVideoCodecs
instead to reduce size:target 'YourTarget' do pod 'AmazonChimeSDKNoVideoCodecs' ... end
-
(Optional) If you want to use background blur and replacement features, add:
target 'YourTarget' do pod 'AmazonChimeSDKMachineLearning' ... end
-
Then run the following command to install pods:
$ pod install --repo-update
-
To open your project, open the newly generated
*.xcworkspace
file in the root directory using XCode. You can do this by issuing the following command in your project folder$ xed .
Note: Do NOT use *.xcodeproj to open project.
-
If you are using background blur and replacement features, under
Build Settings
tab, under theLinking
section, add-framework AmazonChimeSDKMachineLearning
toOther Linker Flags
.
The Amazon Chime SDK is available through SPM. If you don't need video and content share functionality, or software video codec support, you could choose to use no video codecs SPM instead.
-
Open your project in Xcode
-
Go to File > Swift Packages > Add Package Dependency...
-
In the field Enter package repository URL, enter "https://github.com/aws/amazon-chime-sdk-ios-spm". To use no video codecs media module instead, enter "https://github.com/aws/amazon-chime-sdk-ios-no-video-codecs-spm".
-
Enter the latest version(e.g.
0.23.1
) and click Next. -
Choose packages for your project and click Finish.
AmazonChimeSDK
andAmazonChimeSDKMedia
are required. CheckAmazonChimeSDKMachineLearning
if you'd like to use background blur and background replacement.
- Download the
AmazonChimeSDK
andAmazonChimeSDKMedia
binaries from the latest release.- If you'd like to use background blur and background replacement, also download the
AmazonChimeSDKMachineLearning
binary. Otherwise, you can ignore all references toAmazonChimeSDKMachineLearning
in the instructions below. - If you don't need video and content share functionality, or software video codec support, you could use
AmazonChimeSDKMediaNoVideoCodecs
binary instead ofAmazonChimeSDKMedia
and treat all references toAmazonChimeSDKMedia
asAmazonChimeSDKMediaNoVideoCodecs
in the instructions below.
- If you'd like to use background blur and background replacement, also download the
NOTE: We do not support mixing and matching binaries from different releases.
- Unzip and copy the
.framework
s or.xcframework
s toroot
, which depends on which framework your project uses. For Xcode12.3 and later, please use.xcframework
if you have compile issue..xcframework
is available after Amazon Chime SDK iOS v0.15.0
-
Open your
.xcodeproj
file in Xcode and click on your build target. -
Under
Build Settings
tab,-
add
$(PROJECT_DIR)
toFramework Search Path
. -
add
@executable_path/Frameworks
toRunpath Search Paths
. -
under
Linking
section, add the following two flags inOther Linker Flags
:-lc++
-ObjC
-
-
Under
General
tab, look forFrameworks, Libraries, and Embedded Content
section. Click on +, thenAdd Others
, thenAdd Files
.- If you are using traditional
.framework
, specify the location ofAmazonChimeSDK.framework
,AmazonChimeSDKMedia.framework
, andAmazonChimeSDKMachineLearning.framework
from Step 1. If you have compile error while using traditional.framework
, which occurs in Xcode 12.3 and later, please use.xcframework
instead, which is available after Amazon Chime SDK iOS v0.15.0. - If you are using
.xcframework
, specify the location ofAmazonChimeSDK.xcframework
,AmazonChimeSDKMedia.xcframework
, andAmazonChimeSDKMachineLearning.xcframework
from Step 1. - For
AmazonChimeSDK.framework
AmazonChimeSDKMedia.framework
and frameworks, verify thatEmbed & Sign
is selected under theEmbed
option. ForAmazonChimeSDKMachineLearning.framework
, selectDo Not Embed
.
- If you are using traditional
To run the demo application, follow these steps.
git clone git@github.com:aws/amazon-chime-sdk-ios.git
For both targets in /AmazonChimeSDKDemo/Podfile
, replace AMAZON_CHIME_SDK_VERSION
with a specific SDK version, e.g. 0.19.3
or remove it if utilize the latest version of Amazon Chime SDK.
Under /AmazonChimeSDKDemo
, run the following command to install pods:
$ pod install --repo-update
-
Download
AmazonChimeSDKMedia
andAmazonChimeSDKMachineLearning
binaries from the latest release. -
Unzip and copy
AmazonChimeSDKMedia.xcframework
toamazon-chime-sdk-ios/AmazonChimeSDK
folder andAmazonChimeSDKMachineLearning.xcframework
toamazon-chime-sdk-ios/AmazonChimeSDKDemo
folder.
Deploy the serverless demo from amazon-chime-sdk-js
AmazonChimeDemoSDKBroadcast.appex
is a Broadcast Extension for device level screen sharing used by AmazonChimeSDKDemo, verify thatEmbed without Signing
is selected under theEmbed
option. Remove it fromFrameworks, Libraries, and Embedded Content
section if you do not wish to test this.
-
For each target, under
Signing & Capabilities
tab,Signing
section, use your own Apple Developer team and Bundle Identifier.- (Optional)
App Groups
section, select your own app groups if you wish to test sharing device level screen capture. See Content Share for more details.
-
Update server URL and region:
- For Swift demo, update
AppConfiguration.swift
with the server URL and region of the serverless demo. - For ObjC demo, update
ViewControllerObjC.h
with the server URL and region of the serverless demo.
- For Swift demo, update
-
(Optional) Update
broadcastBundleId
andappGroupId
in BOTHAppConfiguration.swift
andSampleHandler.swift
with the broadcast upload extension bundle ID and App Group ID if you want to test sharing device level screen capture. See Content Share for more details.
Open AmazonChimeSDKDemo.xcworkspace
file in AmazonChimeSDKDemo/
using XCode, select the AmazonChimeSDKDemoPods
from the scheme dropdown list in the top bar of Xcode IDE, choose a build device and click the run button.
Open AmazonChimeSDKDemo.xcworkspace
file in AmazonChimeSDKDemo/
using XCode, select the AmazonChimeSDKDemo
from the scheme dropdown list in the top bar of Xcode IDE, choose a build device and click the run button.
On the joining screen, choose to join the meeting without CallKit
or join via CallKit
incoming/outgoing call. Since the demo app does not have Push Notification, it delays joining via incoming call by 10 seconds to give user enough time to background the app or lock the screen to mimic the behavior.
If you discover a potential security issue in this project we ask that you notify AWS/Amazon Security via our vulnerability reporting page. Please do not create a public GitHub issue.
- Starting a session
- Device
- Audio
- Video
- Screen and content share
- Metrics
- Data Message
- Stopping a session
- Amazon Voice Focus
- Custom Video Source
You need to start the meeting session to start sending and receiving audio.
meetingSession.audioVideo.start()
The default configurations are:
- audio format is Stereo/48KHz i.e Stereo Audio with 48KHz sampling rate (stereo48K)
- audio device capabilities are input and output i.e. both microphone and speaker are enabled, which requires audio recording permissions (use
.outputOnly
or.none
to avoid requiring audio recording permissions) - call kit disabled
- audio redundancy enabled
You can specify non-default options in AudioVideoConfiguration
, and then start the meeting session.
var audioVideoConfig = AudioVideoConfiguration(
audioMode: .mono48k,
audioDeviceCapabilities: .outputOnly,
callKitEnabled: true,
enableAudioRedundancy: false,
reconnectTimeoutMs: 180 * 1000)
meetingSession.audioVideo.start(audioVideoConfiguration: audioVideoConfig)
Note: To avoid missing any events, add an observer before the session starts. You can remove the observer by calling meetingSession.audioVideo.removeAudioVideoObserver(observer).
class MyAudioVideoObserver: AudioVideoObserver {
func audioSessionDidStartConnecting(reconnecting: Bool) {
if (reconnecting) {
// e.g. the network connection is dropped.
}
}
func audioSessionDidStart(reconnecting: Bool) {
// Meeting session starts.
// Can use realtime, devices APIs.
}
func audioSessionDidDrop() {}
func audioSessionDidStopWithStatus(sessionStatus: MeetingSessionStatus) {
// See the "Stopping a session" section for details.
}
func audioSessionDidCancelReconnect() {}
func connectionDidRecover() {}
func connectionDidBecomePoor() {}
func videoSessionDidStartConnecting() {}
func videoSessionDidStartWithStatus(sessionStatus: MeetingSessionStatus) {
// Video session starts.
// Can use video APIs.
}
func videoSessionDidStopWithStatus(sessionStatus: MeetingSessionStatus) {}
meetingSession.audioVideo.addAudioVideoObserver(observer: self)
}
List available audio devices for the meeting.
// An list of MediaDevice objects
let audioDevices = meetingSession.audioVideo.listAudioDevices()
for device in audioDevices {
logger.info(msg: "Device type: \(device.type), label: \(device.label)")
}
Note: You should call this after the session started or it’ll be no-op. You should call chooseAudioDevice with one of devices returned from listAudioDevices().
let audioDevices = audioVideo.listAudioDevices()
val device = /* An item from audioDevices */
meetingSession.audioVideo.chooseAudioDevice(mediaDevice: device)
Note: switchCamera() is no-op if you are using custom camera capture source. Please refer to Custom Video for more details.
Switch to use front or back camera on the device, if available.
meetingSession.audioVideo.switchCamera()
Add a DeviceChangeObserver
to receive a callback when a new audio device connects or when an audio device disconnects. audioDeviceDidChange
includes an updated device list.
class MyDeviceChangeObserver: DeviceChangeObserver {
func audioDeviceDidChange(freshAudioDeviceList: [MediaDevice]) {
// A list of updated MediaDevice objects
for device in freshAudioDeviceList {
logger.info(msg: "Device type: \(device.type), label: \(device.label)")
}
}
meetingSession.audioVideo.addDeviceChangeObserver(observer: self)
}
let activeAudioDevice = meetingSession.audioVideo.getActiveAudioDevice()
When joining a meeting, Mono/16KHz, Mono/48KHz and Stereo/48KHz are supported. Stereo/48KHz will be set as the default audio mode if not explicitly specified when starting the audio session.
meetingSession.audioVideo.start() // starts the audio video session with Stereo/48KHz audio, audio input and output devices enabled, callkit disabled, and audio redundancy enabled
meetingSession.audioVideo.start(audioVideoConfiguration) // starts the audio video session with the specified [AudioVideoConfiguration]
Note: So far, you've added observers to receive device and session lifecycle events. In the following use cases, you'll use the real-time API methods to send and receive volume indicators and control mute state.
let muted = meetingSession.audioVideo.realtimeLocalMute() // returns true if muted, false if failed
let unmuted = meetingSession.audioVideo.realtimeLocalUnmute // returns true if unmuted, false if failed
Use case 10. Add an observer to observe realtime events such as volume changes/signal change/muted status of a specific attendee.
You can use this to build real-time indicators UI and get them updated for changes delivered by the array.
Note: These callbacks will only include the delta from the previous callback.
class MyRealtimeObserver: RealtimeObserver {
func volumeDidChange(volumeUpdates: [VolumeUpdate]) {
for currentVolumeUpdate in volumeUpdates {
// Muted, NotSpeaking, Low, Medium, High
logger.info(msg: "\(currentVolumeUpdate.attendeeInfo.attendeeId)'s volume changed: \(currentVolumeUpdate.volumeLevel)")
}
}
func signalStrengthDidChange(signalUpdates: [SignalUpdate]) {
for currentSignalUpdate in signalUpdates {
// None, Low, High
logger.info(msg: "\(currentSignalUpdate.attendeeInfo.attendeeId)'s signal strength changed: \(currentSignalUpdate.signalStrength)")
}
}
func attendeesDidJoin(attendeeInfo: [AttendeeInfo]) {
for currentAttendeeInfo in attendeeInfo {
logger.info(msg: "\(currentAttendeeInfo.attendeeId) joined the meeting")
}
}
func attendeesDidLeave(attendeeInfo: [AttendeeInfo]) {
for currentAttendeeInfo in attendeeInfo {
logger.info(msg: "\(currentAttendeeInfo.attendeeId) left the meeting")
}
}
func attendeesDidDrop(attendeeInfo: [AttendeeInfo]) {
for currentAttendeeInfo in attendeeInfo {
logger.info(msg: "\(currentAttendeeInfo.attendeeId) dropped from the meeting")
}
}
func attendeesDidMute(attendeeInfo: [AttendeeInfo]) {
for currentAttendeeInfo in attendeeInfo {
logger.info(msg: "\(currentAttendeeInfo.attendeeId) muted")
}
}
func attendeesDidUnmute(attendeeInfo: [AttendeeInfo]) {
for currentAttendeeInfo in attendeeInfo {
logger.info(msg: "\(currentAttendeeInfo.attendeeId) unmuted")
}
}
meetingSession.audioVideo.addRealtimeObserver(observer: self)
}
You can use the activeSpeakerDidDetect
event to enlarge or emphasize the most active speaker’s video tile if available. By setting the scoreCallbackIntervalMs
and implementing activeSpeakerScoreDidChange
, you can receive scores of the active speakers periodically.
class MyActiveSpeakerObserver: ActiveSpeakerObserver {
let activeSpeakerObserverId = UUID().uuidString
var observerId: String {
return activeSpeakerObserverId
}
func activeSpeakerDidDetect(attendeeInfo: [AttendeeInfo]) {
if !attendeeInfo.isEmpty {
logger.info(msg: "\(attendeeInfo[0].attendeeId) is the most active speaker")
}
}
var scoresCallbackIntervalMs: Int {
return 1000 // 1 second
}
func activeSpeakerScoreDidChange(scores: [AttendeeInfo: Double]) {
let scoresInString = scores.map { (score) -> String in
let (key, value) = score
return "\(key.attendeeId): \(value)"
}.joined(separator: ",")
logger.info(msg: "Scores of active speakers are: \(scoresInString)")
}
// Calculating the active speaker base on the SDK provided policy, you can provide any custom algorithm
meetingSession.audioVideo.addActiveSpeakerObserver(policy: DefaultActiveSpeakerPolicy(), observer: self)
}
Note: You will need to bind the video to
VideoRenderView
in order to display the video.A local video tile can be identified using
isLocalTile
property.A content video tile can be identified using
isContent
property. See Screen and content share.A tile is created with a new tile ID when the same remote attendee restarts the video.
You can find more details on adding/removing/viewing video from Building a meeting application on ios using the Amazon Chime SDK.
You can call startRemoteVideo
to start receiving remote videos, as this doesn’t happen by default.
meetingSession.audioVideo.startRemoteVideo()
stopRemoteVideo
stops receiving remote videos and triggers onVideoTileRemoved
for existing remote videos.
meetingSession.audioVideo.stopRemoteVideo()
class MyVideoTileObserver: VideoTileObserver {
func videoTileDidAdd(tileState: VideoTileState) {
// Ignore local video (see View local video), content video (see Screen and content share)
if tileState.isLocalTile || tileState.isContent {
return
}
let videoRenderView = /* a VideoRenderView object in your application to show the video */
meetingSession.audioVideo.bind(videoView: videoRenderView, tileId: tileState.tileId)
}
func videoTileDidRemove(tileState: VideoTileState) {
// unbind video view to stop viewing the tile
meetingSession.audioVideo.unbindVideoView(tileId: tileState.tileId)
}
meetingSession.audioVideo.addVideoTileObserver(observer: self)
}
// Use internal camera capture for the local video
meetingSession.audioVideo.startLocalVideo()
// Use internal camera capture and set configuration for the video, e.g. simulcastEnabled, maxBitRateKbps
// If maxBitRateKbps is not set, it will be self adjusted depending on number of users and videos in the meeting
// This can be called multiple times to enable/disable simulcast and adjust video max bit rate on the fly
let localVideoConfig = LocalVideoConfiguration(simulcastEnabled: true, maxBitRateKbps: 600)
meetingSession.audioVideo.startLocalVideo(config: localVideoConfig)
// You can switch camera to change the video input device
meetingSession.audioVideo.switchCamera()
// Or you can inject custom video source for local video, see custom video guide
meetingSession.audioVideo.stopLocalVideo()
Note: The local video should be mirrored. Set VideoRenderView.mirror = true
class MyVideoTileObserver: VideoTileObserver {
func videoTileDidAdd(tileState: VideoTileState) {
if tileState.isLocalTile {
let localVideoView = /* a VideoRenderView object in your application to show the video */
meetingSession.audioVideo.bind(videoView: localVideoView, tileId: tileState.tileId)
}
}
}
func videoTileDidRemove(tileState: VideoTileState) {
// unbind video view to stop viewing the tile
meetingSession.audioVideo.unbindVideoView(tileId: tileState.tileId)
}
meetingSession.audioVideo.addVideoTileObserver(observer: self)
}
For more advanced video tile management, take a look at Video Pagination.
Note: When you or other attendees share content (e.g. screen capture or any other VideoSource object), the content attendee (attendee-id#content) joins the session and shares content as if a regular attendee shares a video.
For example, your attendee ID is "my-id". When you call
meetingSession.audioVideo.startContentShare
, the content attendee "my-id#content" will join the session and share your content.
class MyContentShareObserver: ContentShareObserver {
func contentShareDidStart() {
logger.info(msg: "Content Share has started")
}
func contentShareDidStop(status: ContentShareStatus){
logger.info(msg: "Content Share has stopped")
}
meetingSession.audioVideo.addContentShareObserver(observer: self)
let contentShareSource = /* a ContentShareSource object, can use InAppScreenCaptureSource for screen share or any subclass with custom video source */
// ContentShareSource object is not managed by SDK, builders need to start, stop, release accordingly
meetingSession.audioVideo.startContentShare(source: contentShareSource)
}
You can set configuration for content share, e.g. maxBitRateKbps. Actual quality achieved may vary throughout the call depending on what system and network can provide.
let contentShareConfig = LocalVideoConfiguration(maxBitRateKbps: 200)
meetingSession.audioVideo.startContentShare(source: contentShareSource, config: contentShareConfig)
See Content Share for more details.
meetingSession.audioVideo.stopContentShare()
Chime SDK allows two simultaneous content shares per meeting. Remote content shares will trigger onVideoTileAdded
, while local share will not. To render the video for preview, add a VideoSink
to the VideoSource
in the ContentShareSource
.
class MyVideoTileObserver: VideoTileObserver {
func videoTileDidAdd(tileState: VideoTileState) {
if (tileState.isContent) {
// tileState.attendeeId is formatted as "attendee-id#content"
let attendeeId = tileState.attendeeId
// Get the attendee ID from "attendee-id#content"
let baseAttendeeId = DefaultModality(attendeeId).base()
logger.info(msg: "$baseAttendeeId is sharing screen")
let screenVideoView = /* a VideoRenderView object in your application to show the video */
meetingSession.audioVideo.bindVideoView(videoView: screenVideoView, tileId: tileState.tileId)
}
}
func videoTileDidRemove(tileState: VideoTileState) {
meetingSession.audioVideo.unbindVideoView(tileId: tileState.tileId)
}
meetingSession.audioVideo.addVideoTileObserver(observer: self)
}
See ObservableMetric
for more available metrics and to monitor audio, video, and content share quality.
class MyMetricsObserver: MetricsObserver {
func metricsDidReceive(metrics: [AnyHashable: Any]) {
logger.info(msg: "Media metrics have been received: \(metrics)")
}
meetingSession.audioVideo.addMetricsObserver(observer: self)
}
You can receive real-time messages from multiple topics after starting the meeting session.
class MyDataMessageObserver: DataMessageObserver {
let dataMessageTopic = "chat"
// A throttled message is returned by backend from local sender
func dataMessageDidReceived(dataMessage: DataMessage) {
logger.info(msg: "\(dataMessage.timestampMs) \(dataMessage.text()) \(dataMessage.senderAttendeeId)")
}
// You can also subscribe to multiple topics.
meetingSession.audioVideo.addRealtimeDataMessageObserver(topic: dataMessageTopic, observer: self)
}
You can send real time message to any topic, to which the observers that have subscribed will be notified.
Note: Topic needs to be alpha-numeric and it can include hyphen and underscores. Data cannot exceed 2kb and lifetime is optional but positive integer.
let dataMessageTopic = "chat"
let dataMessageLifetimeMs = 1000
do {
// Send "Hello Chime" to any subscribers who are listening to "chat" topic with 1 seconds of lifetime
try meetingSession
.audioVideo
.realtimeSendDataMessage(topic: dataMessageTopic,
data: "Hello Chime",
lifetimeMs: dataMessageLifetimeMs)
} catch let err as SendDataMessageError {
logger.error(msg: "Failed to send message! \(err)")
} catch {
logger.error(msg: "Unknown error \(error.localizedDescription)")
}
Note: Make sure to remove all the observers and release resources you have added to avoid any memory leaks.
class MyAudioVideoObserver: AudioVideoObserver {
func audioSessionDidStopWithStatus(sessionStatus: MeetingSessionStatus) {
// This is where meeting ended.
// You can do some clean up work here.
}
func videoSessionDidStopWithStatus(sessionStatus: MeetingSessionStatus) {
// This will be invoked as well.
}
meetingSession.audioVideo.addAudioVideoObserver(observer: self)
meetingSession.audioVideo.stop()
}
Amazon Voice Focus reduces the background noise in the meeting for better meeting experience. For more details, see Amazon Voice Focus.
let enabled = audioVideo.realtimeSetVoiceFocusEnabled(enabled: true) // enabling Amazon Voice Focus successful
let disabled = audioVideo.realtimeSetVoiceFocusEnabled(enabled: false) // disabling Amazon Voice Focus successful
Custom video source allows you to control the video, such as applying a video filter. For more details, see Custom Video.
Starting from version 0.23.3, the SDK starts sending redundant audio data to our servers on detecting packet loss to help reduce its effect on audio quality. Redundant audio packets are only sent out for packets containing active audio, i.e. speech or music. This may increase the bandwidth consumed by audio to up to 3 times the normal amount depending on the amount of packet loss detected. The SDK will automatically stop sending redundant data if it hasn't detected any packet loss for 5 minutes.
If you need to disable this feature, you can do so through the AudioVideoConfiguration before starting the session.
meetingSession.audioVideo.start(AudioVideoConfiguration(enableAudioRedundancy: false))
While there is an option to disable the feature, we recommend keeping it enabled for improved audio quality. One possible reason to disable it might be if your customers have very strict bandwidth limitations.
Refer to General FAQ for Amazon Chime SDK.
Applications can get logs from Chime SDK by passing instances of Logger when creating MeetingSession. Amazon Chime SDK has some default implementations of logger that your application can use, such as ConsoleLogger which logs into console. ConsoleLogger
is set to INFO
level as default. Therefore, in order to get all logs, including media logs, create logger by following:
logger = ConsoleLogger(name: "logger", level: .DEFAULT)
You and your end users are responsible for all Content (including any images) uploaded for use with background replacement, and must ensure that such Content does not violate the law, infringe or misappropriate the rights of any third party, or otherwise violate a material term of your agreement with Amazon (including the documentation, the AWS Service Terms, or the Acceptable Use Policy).
Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.