yjjnls/Notes

View on GitHub
gstreamer/src/webrtc/simple/webrtc.c

Summary

Maintainability
Test Coverage
#include <gst/gst.h>
#include <gst/sdp/sdp.h>
#include <gst/webrtc/webrtc.h>

#include <string.h>

static GMainLoop *loop;
static GstElement *pipe1, *webrtc1, *webrtc2;
static GstBus *bus1;

static gboolean _bus_watch(GstBus *bus, GstMessage *msg, GstElement *pipe)
{
    switch (GST_MESSAGE_TYPE(msg))
    {
        case GST_MESSAGE_STATE_CHANGED:
            if (GST_ELEMENT(msg->src) == pipe)
            {
                GstState old, new, pending;

                gst_message_parse_state_changed(msg, &old, &new, &pending);

                {
                    gchar *dump_name = g_strconcat("state_changed-",
                                                   gst_element_state_get_name(old), "_",
                                                   gst_element_state_get_name(new), NULL);
                    GST_DEBUG_BIN_TO_DOT_FILE_WITH_TS(GST_BIN(msg->src),
                                                      GST_DEBUG_GRAPH_SHOW_ALL, dump_name);
                    g_free(dump_name);
                }
            }
            break;
        case GST_MESSAGE_ERROR:
        {
            GError *err = NULL;
            gchar *dbg_info = NULL;

            GST_DEBUG_BIN_TO_DOT_FILE_WITH_TS(GST_BIN(pipe),
                                              GST_DEBUG_GRAPH_SHOW_ALL, "error");

            gst_message_parse_error(msg, &err, &dbg_info);
            g_printerr("ERROR from element %s: %s\n",
                       GST_OBJECT_NAME(msg->src), err->message);
            g_printerr("Debugging info: %s\n", (dbg_info) ? dbg_info : "none");
            g_error_free(err);
            g_free(dbg_info);
            g_main_loop_quit(loop);
            break;
        }
        case GST_MESSAGE_EOS:
        {
            GST_DEBUG_BIN_TO_DOT_FILE_WITH_TS(GST_BIN(pipe),
                                              GST_DEBUG_GRAPH_SHOW_ALL, "eos");
            g_print("EOS received\n");
            g_main_loop_quit(loop);
            break;
        }
        default:
            break;
    }

    return TRUE;
}
#include<stdio.h>
static
GstPadProbeReturn cb_have_data(GstPad *pad, GstPadProbeInfo *info, gpointer user_data)
{
    printf("-");
    return GST_PAD_PROBE_OK;
}
static void _webrtc_pad_added(GstElement *webrtc, GstPad *new_pad, GstElement *pipe)
{
    printf("\n\n\n\n\n\n@@@@@@@@@@@@\n\n\n");
    GstElement *out;
    GstPad *sink;

    if (GST_PAD_DIRECTION(new_pad) != GST_PAD_SRC)
        return;

    out = gst_parse_bin_from_description(
        "rtph264depay name=test ! avdec_h264 ! "
        "videoconvert ! queue ! autovideosink sync=false",
        TRUE, NULL);
    gst_bin_add(GST_BIN(pipe), out);

    GstElement *elem = gst_bin_get_by_name(GST_BIN(out), "test");
    GstPad *pad = gst_element_get_static_pad(elem, "src");
    gst_pad_add_probe(pad, GST_PAD_PROBE_TYPE_BUFFER, cb_have_data, NULL, NULL);
    gst_object_unref(pad);

    gst_element_sync_state_with_parent(out);

    sink = out->sinkpads->data;

    gst_pad_link(new_pad, sink);
}

static void _on_answer_created(GstPromise *promise, gpointer user_data)
{
    GstWebRTCSessionDescription *answer = NULL;
    const GstStructure *reply;
    gchar *desc;

    g_assert(gst_promise_wait(promise) == GST_PROMISE_RESULT_REPLIED);
    reply = gst_promise_get_reply(promise);
    gst_structure_get(reply, "answer", GST_TYPE_WEBRTC_SESSION_DESCRIPTION, &answer, NULL);
    gst_promise_unref(promise);
    desc = gst_sdp_message_as_text(answer->sdp);
    g_print("===========>Webrtc2 created answer:\n%s\n", desc);
    g_free(desc);

    g_signal_emit_by_name(webrtc1, "set-remote-description", answer, NULL);
    g_signal_emit_by_name(webrtc2, "set-local-description", answer, NULL);

    gst_webrtc_session_description_free(answer);
}

static void _on_offer_created(GstPromise *promise, gpointer user_data)
{
    GstWebRTCSessionDescription *offer = NULL;
    const GstStructure *reply;
    gchar *desc;

    g_assert(gst_promise_wait(promise) == GST_PROMISE_RESULT_REPLIED);
    reply = gst_promise_get_reply(promise);
    gst_structure_get(reply, "offer", GST_TYPE_WEBRTC_SESSION_DESCRIPTION, &offer, NULL);
    gst_promise_unref(promise);
    gst_sdp_media_add_attribute((GstSDPMedia *)&g_array_index(offer->sdp->medias, GstSDPMedia, 0), "fmtp", "96 profile-level-id=42e01f");

    desc = gst_sdp_message_as_text(offer->sdp);
    g_print("===========>Webrtc1 created offer:\n%s\n", desc);
    g_free(desc);

    g_signal_emit_by_name(webrtc1, "set-local-description", offer, NULL);
    g_signal_emit_by_name(webrtc2, "set-remote-description", offer, NULL);

    promise = gst_promise_new_with_change_func(_on_answer_created, user_data, NULL);
    g_signal_emit_by_name(webrtc2, "create-answer", NULL, promise);

    gst_webrtc_session_description_free(offer);
}
// When the pipeline goes to PLAYING, the on_negotiation_needed() callback will be called, 
// and we will ask webrtcbin to create an offer which will match the pipeline above.
static void _on_negotiation_needed(GstElement *element, gpointer user_data)
{
    GstPromise *promise;

    promise = gst_promise_new_with_change_func(_on_offer_created, user_data, NULL);
    g_signal_emit_by_name(webrtc1, "create-offer", NULL, promise);
}

static void _on_ice_candidate(GstElement *webrtc, guint mlineindex,
                              gchar *candidate, GstElement *other)
{
    g_signal_emit_by_name(other, "add-ice-candidate", mlineindex, candidate);
}

int main(int argc, char *argv[])
{
    gst_init(&argc, &argv);

    loop = g_main_loop_new(NULL, FALSE);
    pipe1 = gst_parse_launch(
        "rtspsrc location=rtsp://172.16.66.65/id=1 ! rtph264depay ! queue ! rtph264pay config-interval=-1 ! queue ! "
        "application/x-rtp,media=video,payload=96,encoding-name=H264 ! "
        "webrtcbin name=send webrtcbin name=recv ! rtph264depay name=test ! avdec_h264 ! videoconvert ! queue ! autovideosink sync=false",
        NULL);
    bus1 = gst_pipeline_get_bus(GST_PIPELINE(pipe1));
    gst_bus_add_watch(bus1, (GstBusFunc)_bus_watch, pipe1);
    //  This is the gstwebrtc entry point where we create the offer.
    //  It will be called when the pipeline goes to PLAYING.
    webrtc1 = gst_bin_get_by_name(GST_BIN(pipe1), "send");
    g_signal_connect(webrtc1, "on-negotiation-needed", G_CALLBACK(_on_negotiation_needed), NULL);
    // Incoming streams will be exposed via this signal
    webrtc2 = gst_bin_get_by_name(GST_BIN(pipe1), "recv");
    // g_signal_connect(webrtc2, "pad-added", G_CALLBACK(_webrtc_pad_added), pipe1);
    // transfer the candidates of the two webrtcs
    g_signal_connect(webrtc1, "on-ice-candidate", G_CALLBACK(_on_ice_candidate), webrtc2);
    g_signal_connect(webrtc2, "on-ice-candidate", G_CALLBACK(_on_ice_candidate), webrtc1);

    g_print("===========>Starting pipeline<==========\n");
    gst_element_set_state(GST_ELEMENT(pipe1), GST_STATE_PLAYING);

    g_main_loop_run(loop);

    gst_element_set_state(GST_ELEMENT(pipe1), GST_STATE_NULL);
    g_print("===========>Pipeline stopped<==========\n");

    gst_object_unref(webrtc1);
    gst_object_unref(webrtc2);
    gst_bus_remove_watch(bus1);
    gst_object_unref(bus1);
    gst_object_unref(pipe1);

    gst_deinit();

    return 0;
}