[Live-devel] Streaming live AMR and H263P
salick at videocells.com
salick at videocells.com
Wed Mar 29 04:33:26 PST 2006
> Because your encoding application delivers discrete (H.263 video and
> AMR audio) frames, then your "FramedSource" subclasses - that you
> will write - should also deliver discrete frames. (Don't just
> concatenate the encoded frames into a buffer, otherwise you'll have
> to parse the data into frames once again, before they can be streamed.)
>
> Your H.263 video and AMR audio "FramedSource" subclasses will feed
> into a "H263plusVideoRTPSink" and a "AMRAudioRTPSink" (not a
> "*Source"), respectively.
For AMR source I sort of get the implementation I need to do, by looking
at the AMRFileSource. I can see it sets the fTo byte array as the frame,
fPresentationTime,fDurationInMicroseconds... and then calls nextTask().
But there's no filesource implementation for H.263P, so How should I
implement the doGetNextFrame() suppose I have one H.263 frame in a byte
array (lets call it frameBuffer).
More information about the live-devel
mailing list