iPadOS 14.5 BETA
Not sure which build I'm using.
When the mediaRecorder is paused and then requestData() is called, the onDataAvailable event is fired, but only 1117 bytes are returned (header???). This is true even if requestData() is called inside a setTimeout with a 10 sec delay. The recorded media can not be played until a second call to requestData() is made.
requestData should "raise a dataavailable event containing a Blob object of the captured media as it was when the method was called", not just the first 1117 bytes.
1) click Start Camera
2) Click Start Recording
3) Click Pause. In the console, notice that event.data size is 1117
4) Click Play. Notice that there is no video to play in the video element. In the console, the current blob array only contains one blob with 1117 bytes. Also note that the play button called requestData() again and this time the dataavailable event fired with the rest of the recorded data.
5) Click Play again. Now the recorded video plays. In the console, the current blob array now has two blobs.
Created attachment 422682 [details]
Committed r274154: <https://commits.webkit.org/r274154>
All reviewed patches have been landed. Closing bug and clearing flags on attachment 422682 [details].