Hi could someone please explain to me how to achieve a lipsync between two RTP streams (from the same RTSP session). I'm trying to calculate proper pts for ffmpeg AVPacket but I'm missing something and I can't get my head around this.
I have the following data available to me:
u64RTCP_NTP_TS - NTP timestamp from RTCP Sender Report
u32RTCP_TS - Timestamp from the RTCP Sender Report
u32AudioRTP_TS - Timestamp from the Audio RTP packet
u32VideoRTP_TS - Timestamp from the Video RTP Packet
I've searched for an answer but still I can't get a clear picture of how this should be achieved calculation-wise, what am I still missing.