I can understand how the audio is stretched by CTL-Dragging the edge of an event.  But how is the video "stretched?"  Particularly when quantize to frames is being used. I noticed that the stretching is accomplished in discrete steps, quanta if you wish.  But what is happening within the event?  Are duplicate video frames being added to the video?  Are they evenly distributed?  Or are intermediate frames generated that are a blend of adjacent frames?  If so, what are the parameters employed?  Does anyone know?