Are there any additional costs involved in using your software?

There are no fees you need to pay! But you do need to get your own server as I can’t guarantee that my test servers are always running. Check out the questions about server support for more.

What servers do I need and why do I need them?

WebRTC connects users directly to each other thus allowing you to reduce server load. But you still need a server / multiple server applications to initialize the connection between the users.

Signaling server: This is a node.js server developed by me. It is using (secure) websockets to allow two users to find each other online using any kind of string as an address. For example if user B wants to connect to user A then user A needs to request an address from this server. User B can use that address to initialize a connection to User A.
The java script source code is part of the WebRTC Network and Video Chat package (see server.zip + readme.txt!). I also provide a test server which set up by default.

STUN server: One of the two kinds of ice server. This server is part of the native WebRTC implementation and is needed to establish connections trough a router / NAT. As most users are behind a  router / NAT you will generally need it for any connections that aren’t inside your own LAN or WIFI. Currently, the library uses the server “stun:because-why-not.com:12779” but there are free alternatives available for example googles public stun server “stun:stun.l.google.com:19302”. Note that this server isn’t available in China.  TURN servers such as the application “coturn” support STUN as well. So you won’t need two different server applications if you plan to set up your own.

TURN server: Second kind of ice server. If the STUN server fails to connect two peers directly to each other a TURN server can be used to forward messages between two users. A TURN server is not provided as it might cause a lot of traffic. This means that using the default configuration some users might not be able to connect if their router / NAT configuration doesn’t allow direct connections.
Please send an e-mail with your Unity Asset Store invoice number if you need a TURN server for testing purposes.
A TURN server often supports STUN as well. In this case you only need one. TURN will only be used if STUN failed.

General guidelines:

  • If you are developing / testing only or work on a hobby project: Use the provided test servers
  • If your app runs in a local network without internet: Set up a signalling server within the LAN, no STUN or TURN needed
  • If your app should connect via internet at low cost: Set up a signalling server + use a free STUN server or set-up your own STUN server if you want to make sure it is always available. Some users might not be able to connect without TURN but even many AAA games still use this configuration.
  • If you want to make sure everyone can connect or your application will often be used in businesses (they often block direct connections): Set up a signalling server + one or multiple TURN servers.

Why are there multiple ice servers / which one is used?

WebRTC uses 3 different  ways of establishing a connection

  • host – a connection that connects directly to the other side using the local available IPs
  • srflx – a connection using STUN to avoid problems with router / NAT that might be in between to two systems trying to connect
  • relay – using TURN.  This is the only type that uses an external server to relay data.  This only makes sense if the two other ways failed

WebRTC  / ICE has an internal process to determine which one is best to be used using a priority system (RFC5245) . Usually, this means direct host connections will be tried first, then stun, then turn. It will also prefer udp over tcp. The detail of the process is left to the internals of WebRTC and isn’t determined by the asset.  It actually changes regularly during WebRTC updates.

How to set up my own STUN / TURN server?

The project coturn is a great solution! It acts as a TURN and STUN server.  Check out the INSTALL and readme files of the project to get started.

After installation the configuration is usually done via the file “/etc/turnserver.conf”. For testing with webrtc make sure to set at least the properties “user=myusername:mypassword” and “realm=mydomain.com”.  You can also use a custom port using “listening-port=12779”.  If you use cloud services (like azure) then your server might be behind a NAT/Firewall. In this case you also need to set the property “external-ip”.

After the configuration is done you can use your server using the credentials and the url “turn:mydomain.com:12779”.

To test if your server works properly with webrtc you can use this tool trickle-ice. The results should look like this:

The important lines are of the type “strflx” and “relay”. “srflx” is the systems public IP detected by STUN allowing direct connections trough firewall/NAT. And relay is a IP/ port provided by TURN allowing indirect connections via the server. The type “host” stands for possible direct connections in your own LAN without the help of TURN/STUN!

The node.js signalling server doesn’t work / prints the error message “Invalid message received: ” with some cryptic characters. How to fix this?

This is likely caused by an old node.js version being used. You can use the command “node -v” to find out the version. The server was tested and developed with node.js version 6.9!  Versions with a leading zero e.g. 0.10.x won’t work.

The signaling server fails to listen on the standard ports 80 / 443 for ws/ wss

Usually, these ports require to use root access but using nodejs as root isn’t particular secure. Instead, you can use following command to give nodejs access to ports below 1024:

#if not installed yet: sudo apt-get install libcap2-bin
sudo setcap cap_net_bind_service=+ep `readlink -f \`which node\``

What are those ssl.key / ssl.crt files coming with the signaling server / How to use secure connections?

They contain the SSL certificate needed for the encrypted websocket connections (wss). These files will work for with the native version only (because it doesn’t check the validity of the certificate yet). It will not work with browsers. You will have to make sure  you get a valid certificate and replace the files yourself. One way to get a free certificate is https://letsencrypt.org/.

Why can’t I connect (via the internet)?

For this question there is sadly no easy answer. There are many reasons why connections can fail and the network protocols won’t return any information except “I can’t connect”.
So far most common problems are:

  • Firewall blocks access. Either local or one in a router
  • Router/NAT doesn’t support or allow opening incoming ports and no TURN server is used
  • STUN / TURN URL’s are wrong
  • NetworkConfig is wrong, was never set up or was set up after the Call was already created

If you can’t quickly identify the problem then please try the following steps:

1. Create an empty project with the newest version of the Asset.
Ideally, use only the Unity Editor on Windows 64 bit. (If Android is needed make sure you set up AndroidManifest.xml!)
Open the example scene. Either callscene (CallApp) for the audio/video version or chatscene (ChatApp) for the network version.

2. Start the example app on your test system to connect locally.
The scenes contain two instances of the same app. Enter a test address/room name e.g. test123 and press join (CallApp) or Open Room (ChatApp).
You should now see “Waiting for incoming call address: test123″(CallApp) or “Server started. Address: test123” in the text area or log.
If this doesn’t happen / you see an error: This means the connection to the signaling server failed. This server is responsible to keep track of addresses and relays connection information to then later establish a direct connection)

This could mean the following:

  • Firewall blocked access to server
  • You are connected to the internet via a proxy (which isn’t supported)
  • The internet connection is dead or buggy
    • try “ping google.com”. If you see errors / dropped packages  or a high value for time (e.g. above 400ms) then something is wrong with your internet connect (or you are in China).
  • The signaling test server is dead
    • try “ping because-why-not.com”. If you see a high latency here but low latency with google the test server might be updating or dying. If the problem persists after a few hours send an email via (BecauseWhyNotHelp@gmail.com)
    • Try it with a different internet connection. If the problem persists and you excluded all other problems: send an email.

If it worked close the app and test it on your second system using the same internet connection but a different address / room name. If it fails only on one system is very likely a local problem e.g. a local firewall.

3. Connect two systems that are in the same LAN / Wifi
If this doesn’t work but everything in step 2 worked then it could be due to local network restrictions e.g. Schools and Universities often block direct connections between systems. If you can exclude these problems then please send an email.

4. Connect two systems via internet
Please make sure to read “What servers do I need and why do I need them?” to find out what servers are involved and what they are for.

If all steps before worked but this one fails it could be due to the following reasons:

  • The STUN/TURN server is offline. This server is suppose to open the port of the router to allow incoming connections.  You can find the server URL via the Unity Inspector if you select the Chat/CallApp GameObject (value uIceServer and uIceServer2). Usually something like this: “stun:because-why-not.com:12779″ or turn:because-why-not.com:12779” or “stun:stun.l.google.com:19302” (free google server, doesn’t work in China).  For information how to check if the server works see “How to set up my own STUN / TURN server?”.
  • Both involved routers / NAT’s  don’t allow opening incoming ports automatically. In this case two users can’t connect directly but both might still be able to connect to other users that have a different router. You might need a TURN server to avoid this problem. Please send an e-mail with your Unity Asset Store invoice number if you need a TURN server for testing purposes.  Username and password of this server will change regularly to avoid abuse / too much traffic.
  • A Firewall blocks a wide range of ports / blocks custom protocols: This often happens in big organizations, universities and schools. This problem can’t be solved in a reliable way but in some instances using a VPN might help.

Conferences / multiple users in a single video chat?

This is possible since version 0.97! For this you need to configure the server url to allow shared addresses (see example config file) and set on the client side NetworkConfig.IsConference to true. This will allow multiple connections to listen to the same address for incoming connections creating a full mesh topology between every client calling ICall.Listen(address). e.g. If a new user joins using the Listen method all existing users will receive a new incoming call and the new user will receive for every existing user incoming call events as well! I will publish an example for this use case in the future!

Can I run the signaling server on one of the clients?

Yes, this is possible if all clients are within the same LAN. In this case no ICE server is needed.

How do I run the signaling server on my webserver / apache / php webserver?

If your hosting provider supports node.js you might be able to run it.  Some providers require you to use a specific port or replace the port with an environment variable so it can be set by your provider. If only apache or php is supported you won’t be able to run the signaling server at all.
Please note that I can’t help with provider specific issues.

Can I use WebRTC Video Chat to broadcast my game screen / window?

Currently, only connected webcams are supported as video source! Due to multiple emails asking for this feature I decided to add custom video sources in the future. This will allow you to deliver any kind of image(s) / textures to the Video Chat plugin.

What is the maximum resolution, FPS?

This all depends entirely on your device and the native WebRTC implementation. I recommend allowing the user to decide this so they can optimize it for their system.  Not every device can handle the realtime encoding and decoding of full HD streams even if it has a HD camera connected to it. I consider 320 x 240 and 640 x 480 reasonable fast to use as default value. Have a look at the documentation of the class MediaConfig to see how to configure the resolution and fps.

How many users can I have in a single conference call?

There is no clear limit to this. You can connect as many users as you want to but that doesn’t mean their system and network connection is able to handle it. The asset will fundamentally work on a peer to peer basis. A conference call between 10 users means that each users need to download & decode 9 video and audio streams while also uploading & encoding 9 video and audio streams.  This can work with a low resolution but will likely fail with 1080p even on modern systems. Also note that WebRTC will attempt to reduce the quality if it detects a network or CPU overload.

In the end the only way to determine the maximum number is to test it with your specific setup and use case. For this reason all sample application in the asset support being duplicated multiple times in a single scene and on a single system. This way you can easily similate specific use cases without the need of 10 different devices.

How to set the resolution?

You can have a look at the CallApp.CreateMediaConfig. There you can set a minimal, maximal and ideal resolution for your use case. It isn’t possible to enforce a specific resolution though due to differences in browsers, operation systems and video devices.

How do I change the microphone / speaker (Windows)?

The system uses the default microphone/speaker currently.  In windows this is the device set as “Default Communication Device”. (Right click onto the audio symbol -> Choose “Recording devices” for Microphone or “Playback devices” for speaker ->Right click on the device you want to use -> choose Default Communication Device)

microphone1 microphone2


Why is the video so grainy / pixelated?

The video is set by default to a low resolution. This is needed as some devices have very high quality cameras but not the CPU power to encode videos at such high quality. You can change the resolution in the CallApp.cs example via the method “CreateDefaultConfig” mediaConfig.IdealWidth / mediaConfig.IdeaHeight.

Can I send data / stream between different platforms?

Yes, all platforms supported can connect to each other.

Can I connect the asset to another WebRTC application / library?

No, this is currently not possible. WebRTC doesn’t standardize how a connection is established thus different WebRTC applications are usually not compatible to each other. A future version might allow connecting to other WebRTC based systems but support for this feature will be limited.

Can I stream the unity screen?

By default this is not supported. It is possible to send raw ARGB images via an extension you can request via e-mail (currently in testing).  You could use Unitys RenderTextures to get the raw image data from the frame buffer and then use the feature to send it.

Can I send files?

This functionality isn’t part of the API yet but you can always send byte[] of it yourself using WebRTC Network.

Where is the AudioSource / How can I access the raw PCM data?

There is no AudioSource and raw PCM data access isn’t supported. The native WebRTC Version / the browser will replay the audio automatically while receiving a stream.

How to I switch cameras during the call?

This isn’t possible yet due to restrictions in the C++ wrapper. It is planed to change this in the future though.

Platform specific issues


HTTPS / WSS problems

To connect two peers websockets are used to exchange connection information. Some problems can arise due to restrictions imposed by browsers on websockets. To avoid them ideally host your unity application on https and your signaling server on wss  using the same domain and the same certificate.

Problems you might run into:

  • Browsers might refuse access to WebRTC if you use insecure connections. Especially combining https hosting and using a not secure websocket will not work everywhere!
  • Self signed certificates might not work depending on your browser and configuration! If you do use them make sure you visit the page via https:// first to accept the certificate and then start your webgl app,
  • Firefox (and maybe others) don’t seem to automatically accept certificates if they are needed for Websockets only but not for https. For example if you test your webgl application locally but using the default signaling server configuration on because-why-not.com then Firefox might refuse to connect. The connection will work after visiting https://because-why-not.com once, thus allowing firefox to accept and store the certificate. The problem might reoccur after my server updates its certificate(happens every few months)!

The UI is all black

This is a unity bug. Close unity, delete the library folder and then start again.


Unity Editor freezes while connecting / stopping the editor while connecting

Using ssl (due to the wss websocket connection for the signaling server) can cause unity to freeze. This is a bug in unity / mono. It is announced to be fixed in version 5.3.7. Until then I added a workaround that should reduce freezing to 5-20 seconds max instead of the usual 5 minutes. This shouldn’t happen anymore in more recent versions. If it does happen please create a test project and send an error report.


Why can’t I connect on Android? Why is audio / video failing on Android?

A common problem for Android users is a missing AndroidManifest.xml file or it is simple placed in the wrong location. This file is project specific and can’t be supplied/ placed by the Asset. It is needed to give your application the rights to access internet, video and audio. All needed for WebRTC to run properly. Please check the readme.txt for how to set up the project for Android and what permissions need to be requested for the asset to work. There is also an example file in WebRtcNetwork\Plugins\android\AndroidManifest.xml.
The related unity documentation can be found here.  Please make sure to read the documentation and place your file at the exact location: Assets/Plugins/Android/AndroidManifest.xml !

You can ensure your properties are set correctly in the final apk via the command “%YOUR_ANDROID_SDK%/build-tools\23.0.2\aapt.exe l -a your_android_app.apk”

Why can’t I hear the other side speaking on Android / why is the audio so quite?

WebRTC Android doesn’t treat the audio received from the other peer as “Media”.  Instead it will treat it like a phone call. It will automatically set the volume to a level that is comfortable for someone who holds the phone directly onto the ear. The file AndroidHelper.cs (V0.975) contains multiple functions that can help you

  • AndroidHelper.SetSpeakerOn(true) – Turns on the speaker of the phone.
  • ICall.SetVolume – Android applications support a volume above 1 (up to 10).
  • AndroidHelper.SetCommunication() – This switches Android in communication mode. The user can then use the audio buttons to change the call volume.


Leave a Reply