1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
|
/*
Copyright (C) 2015-2016, 2019 CodeWeavers, Inc
This library is free software; you can redistribute it and/or
modify it under the terms of the GNU Lesser General Public
License as published by the Free Software Foundation; either
version 2.1 of the License, or (at your option) any later version.
This library is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
License along with this library; if not, see <http://www.gnu.org/licenses/>.
*/
#include "config.h"
#include "spice-client.h"
#include "spice-common.h"
#include "spice-channel-priv.h"
#include "common/recorder.h"
#include "channel-display-priv.h"
#include <gst/gst.h>
#include <gst/app/gstappsrc.h>
#include <gst/app/gstappsink.h>
#include <gst/video/gstvideometa.h>
/* -GST_EVENT_QOS msg
*
* This is QOS measured by gstreamer for every buffer in order to let pipeline elements know the real-time performance
* and adjust accordingly if possible. For example it may cause frame drops (any adjustment should be notified by "Received QOS MSG" msg)
* (QOS+SYNC has to be TRUE)
*
* proportion: running avg of the ratio between synced presenting times and frame rate
* diff: jitter, presenting time difference from frame PTS.
* timestamp:
*
* more info:
* https://gstreamer.freedesktop.org/documentation/plugin-development/advanced/qos.html
*
*
*
*
* -BUFFERS QOS
*
* This is alternative measurements are not part of gstreamer QOS system but partly similar
*
* queue: number of frames which pushed to gstreamer pipeline and did not arrive to sink yet. (some elements may have their own queues
* so that queue will never be 1 but you can monitor it is not accumulating - accumulating means client is not keeping up)
* rate: Similar idea to proportion but a bit different- running avg of the ratio between buffers sink arrival times difference and framerate
* - As value goes more higher above 1 means pipeline processing takes too long for current frame rate.
*
*
*
*
*
* -Feedback QOS
*
* If pipeline element changing its behaviour because of GST_EVENT_QOS should be notified by this msg.
*
*
*
*
* -GST_EVENT_LATENCY
*
* Estimation of the pipeline latency in the beginning (if LIVE is TRUE) so that frames can be synchronized correctly
*
*
*/
#define QOS FALSE // TRUE will emit qos events upstream which may cause decoder to drop frames (should be effective only with SYNC)
#define SYNC FALSE // TRUE will sync buffer times with clock time
#define LIVE FALSE // TRUE will add pipeline latency estimation (should be effective only with SYNC)
#define DROP FALSE // TRUE if you want pipeline to decide to drop frames
GstClockTime last;
typedef struct SpiceGstFrame SpiceGstFrame;
/* GStreamer decoder implementation */
#if GST_CHECK_VERSION(1,14,0)
static GstStaticCaps stream_reference = GST_STATIC_CAPS("timestamp/spice-stream");
#endif
typedef struct SpiceGstDecoder {
VideoDecoder base;
/* ---------- GStreamer pipeline ---------- */
GstAppSrc *appsrc;
GstAppSink *appsink;
GstElement *pipeline;
GstClock *clock;
/* ---------- Decoding and display queues ---------- */
uint32_t last_mm_time;
gdouble avg_rate;
GMutex queues_mutex;
GQueue *decoding_queue;
SpiceGstFrame *display_frame;
guint timer_id;
guint pending_samples;
guint queue; // queue may not be accurate if qos or drop is true
} SpiceGstDecoder;
#define VALID_VIDEO_CODEC_TYPE(codec) \
(codec > 0 && codec < G_N_ELEMENTS(gst_opts))
/* Decoded frames are big so limit how many are queued by GStreamer */
#define MAX_DECODED_FRAMES 2
#define DO_RUNNING_AVG(avg,val,size) (((val) + ((size)-1) * (avg)) / (size))
/* generic running average, this has a neutral window size */
#define UPDATE_RUNNING_AVG(avg,val) DO_RUNNING_AVG(avg,val,8)
/* the windows for these running averages are experimentally obtained.
* positive values get averaged more while negative values use a small
* window so we can react faster to badness. */
#define UPDATE_RUNNING_AVG_P(avg,val) DO_RUNNING_AVG(avg,val,16)
#define UPDATE_RUNNING_AVG_N(avg,val) DO_RUNNING_AVG(avg,val,4)
/* GstPlayFlags enum is in plugin's header which should not be exported.
* https://bugzilla.gnome.org/show_bug.cgi?id=784279
*/
typedef enum {
GST_PLAY_FLAG_VIDEO = (1 << 0),
GST_PLAY_FLAG_AUDIO = (1 << 1),
GST_PLAY_FLAG_TEXT = (1 << 2),
GST_PLAY_FLAG_VIS = (1 << 3),
GST_PLAY_FLAG_SOFT_VOLUME = (1 << 4),
GST_PLAY_FLAG_NATIVE_AUDIO = (1 << 5),
GST_PLAY_FLAG_NATIVE_VIDEO = (1 << 6),
GST_PLAY_FLAG_DOWNLOAD = (1 << 7),
GST_PLAY_FLAG_BUFFERING = (1 << 8),
GST_PLAY_FLAG_DEINTERLACE = (1 << 9),
GST_PLAY_FLAG_SOFT_COLORBALANCE = (1 << 10),
GST_PLAY_FLAG_FORCE_FILTERS = (1 << 11),
} SpiceGstPlayFlags;
/* ---------- SpiceGstFrame ---------- */
struct SpiceGstFrame {
GstClockTime timestamp;
GstBuffer *encoded_buffer;
SpiceFrame *encoded_frame;
GstSample *decoded_sample;
guint queue_len;
};
static SpiceGstFrame *create_gst_frame(GstBuffer *buffer, SpiceFrame *frame)
{
SpiceGstFrame *gstframe = g_new(SpiceGstFrame, 1);
gstframe->timestamp = GST_BUFFER_PTS(buffer);
#if GST_CHECK_VERSION(1,14,0)
GstReferenceTimestampMeta *time_meta;
time_meta = gst_buffer_get_reference_timestamp_meta(buffer, gst_static_caps_get(&stream_reference));
if (time_meta) {
gstframe->timestamp = time_meta->timestamp;
}
#endif
gstframe->encoded_buffer = gst_buffer_ref(buffer);
gstframe->encoded_frame = frame;
gstframe->decoded_sample = NULL;
return gstframe;
}
static void free_gst_frame(SpiceGstFrame *gstframe)
{
gst_buffer_unref(gstframe->encoded_buffer);
// encoded_frame was owned by encoded_buffer, don't release it
g_clear_pointer(&gstframe->decoded_sample, gst_sample_unref);
g_free(gstframe);
}
/* ---------- GStreamer pipeline ---------- */
static void schedule_frame(SpiceGstDecoder *decoder);
RECORDER(frames_stats, 64, "Frames statistics");
static int spice_gst_buffer_get_stride(GstBuffer *buffer)
{
GstVideoMeta *video = gst_buffer_get_video_meta(buffer);
return video && video->n_planes > 0 ? video->stride[0] : SPICE_UNKNOWN_STRIDE;
}
/* main context */
static gboolean display_frame(gpointer video_decoder)
{
SpiceGstDecoder *decoder = (SpiceGstDecoder*)video_decoder;
SpiceGstFrame *gstframe;
GstCaps *caps;
gint width, height;
GstStructure *s;
GstBuffer *buffer;
GstMapInfo mapinfo;
g_mutex_lock(&decoder->queues_mutex);
decoder->timer_id = 0;
gstframe = decoder->display_frame;
decoder->display_frame = NULL;
g_mutex_unlock(&decoder->queues_mutex);
/* If the queue is empty we don't even need to reschedule */
g_return_val_if_fail(gstframe, G_SOURCE_REMOVE);
if (!gstframe->decoded_sample) {
spice_warning("got a frame without a sample!");
goto error;
}
caps = gst_sample_get_caps(gstframe->decoded_sample);
if (!caps) {
spice_warning("GStreamer error: could not get the caps of the sample");
goto error;
}
s = gst_caps_get_structure(caps, 0);
if (!gst_structure_get_int(s, "width", &width) ||
!gst_structure_get_int(s, "height", &height)) {
spice_warning("GStreamer error: could not get the size of the frame");
goto error;
}
buffer = gst_sample_get_buffer(gstframe->decoded_sample);
if (!gst_buffer_map(buffer, &mapinfo, GST_MAP_READ)) {
spice_warning("GStreamer error: could not map the buffer");
goto error;
}
stream_display_frame(decoder->base.stream, gstframe->encoded_frame,
width, height, spice_gst_buffer_get_stride(buffer), mapinfo.data);
gst_buffer_unmap(buffer, &mapinfo);
error:
free_gst_frame(gstframe);
schedule_frame(decoder);
return G_SOURCE_REMOVE;
}
/* Returns the decoding queue entry that matches the specified GStreamer buffer.
*
* The entry is identified based on the buffer timestamp. However sometimes
* GStreamer returns the same buffer twice (and the second time the entry may
* have been removed already) or buffers that have a modified, and thus
* unrecognizable, timestamp. In such cases NULL is returned.
*
* queues_mutex must be held.
*/
static GList *find_frame_entry(SpiceGstDecoder *decoder, GstBuffer *buffer)
{
GstClockTime buffer_ts = GST_BUFFER_PTS(buffer);
#if GST_CHECK_VERSION(1,14,0)
GstReferenceTimestampMeta *time_meta;
time_meta = gst_buffer_get_reference_timestamp_meta(buffer, gst_static_caps_get(&stream_reference));
if (time_meta) {
buffer_ts = time_meta->timestamp;
}
#endif
GList *l = g_queue_peek_head_link(decoder->decoding_queue);
while (l) {
const SpiceGstFrame *gstframe = l->data;
if (gstframe->timestamp == buffer_ts) {
return l;
}
l = l->next;
}
return NULL;
}
/* Pops the queued frames up to and including the specified frame.
* All frames are freed except that last frame which belongs to the caller.
* Returns the number of freed frames.
*
* queues_mutex must be held.
*/
static guint32 pop_up_to_frame(SpiceGstDecoder *decoder, const SpiceGstFrame *popframe)
{
SpiceGstFrame *gstframe;
guint32 freed = 0;
while ((gstframe = g_queue_pop_head(decoder->decoding_queue)) != popframe) {
free_gst_frame(gstframe);
freed++;
}
return freed;
}
/* Helper for schedule_frame().
*
* queues_mutex must be held.
*/
static void fetch_pending_sample(SpiceGstDecoder *decoder)
{
GstSample *sample = gst_app_sink_pull_sample(decoder->appsink);
if (sample) {
// account for the fetched sample
decoder->pending_samples--;
GstBuffer *buffer = gst_sample_get_buffer(sample);
/* gst_app_sink_pull_sample() sometimes returns the same buffer twice
* or buffers that have a modified, and thus unrecognizable, PTS.
* Blindly removing frames from the decoding_queue until we find a
* match would only empty the queue, resulting in later buffers not
* finding a match either, etc. So check the buffer has a matching
* frame first.
*/
GList *l = find_frame_entry(decoder, buffer);
if (l) {
SpiceGstFrame *gstframe = l->data;
/* Dequeue this and any dropped frames */
guint32 dropped = pop_up_to_frame(decoder, gstframe);
if (dropped) {
SPICE_DEBUG("the GStreamer pipeline dropped %u frames", dropped);
}
/* The frame is now ready for display */
gstframe->decoded_sample = sample;
decoder->display_frame = gstframe;
} else {
spice_warning("got an unexpected decoded buffer!");
gst_sample_unref(sample);
}
} else {
// no more samples to get, possibly some sample was dropped
decoder->pending_samples = 0;
spice_warning("GStreamer error: could not pull sample");
}
}
/* main loop or GStreamer streaming thread */
static void schedule_frame(SpiceGstDecoder *decoder)
{
guint32 now = stream_get_time(decoder->base.stream);
g_mutex_lock(&decoder->queues_mutex);
while (!decoder->timer_id) {
while (decoder->display_frame == NULL && decoder->pending_samples) {
fetch_pending_sample(decoder);
}
SpiceGstFrame *gstframe = decoder->display_frame;
if (!gstframe) {
break;
}
if (spice_mmtime_diff(gstframe->encoded_frame->mm_time, now) >= 0) {
decoder->timer_id = g_timeout_add(gstframe->encoded_frame->mm_time - now,
display_frame, decoder);
} else if (decoder->display_frame && !decoder->pending_samples) {
/* Still attempt to display the least out of date frame so the
* video is not completely frozen for an extended period of time.
*/
decoder->timer_id = g_timeout_add(0, display_frame, decoder);
} else {
SPICE_DEBUG("%s: rendering too late by %u ms (ts: %u, mmtime: %u), dropping",
__FUNCTION__, now - gstframe->encoded_frame->mm_time,
gstframe->encoded_frame->mm_time, now);
stream_dropped_frame_on_playback(decoder->base.stream);
decoder->display_frame = NULL;
free_gst_frame(gstframe);
}
}
g_mutex_unlock(&decoder->queues_mutex);
}
/* GStreamer thread
*
* Decoded frames are big so we rely on GStreamer to limit how many are
* buffered (see MAX_DECODED_FRAMES). This means we must not pull the samples
* as soon as they become available. Instead just increment pending_samples so
* schedule_frame() knows whether it can pull a new sample when it needs one.
*
* Note that GStreamer's signals are not always run in the main context, hence
* the schedule_frame() + display_frame() mechanism. So we might as well use
* a callback here (lower overhead).
*/
static GstFlowReturn new_sample(GstAppSink *gstappsink, gpointer video_decoder)
{
SpiceGstDecoder *decoder = video_decoder;
g_mutex_lock(&decoder->queues_mutex);
decoder->pending_samples++;
if (decoder->timer_id && decoder->display_frame) {
g_mutex_unlock(&decoder->queues_mutex);
return GST_FLOW_OK;
}
g_mutex_unlock(&decoder->queues_mutex);
schedule_frame(decoder);
return GST_FLOW_OK;
}
static void free_pipeline(SpiceGstDecoder *decoder)
{
if (!decoder->pipeline) {
return;
}
gst_element_set_state(decoder->pipeline, GST_STATE_NULL);
gst_object_unref(decoder->appsrc);
if (decoder->appsink) {
gst_object_unref(decoder->appsink);
}
gst_object_unref(decoder->pipeline);
gst_object_unref(decoder->clock);
decoder->pipeline = NULL;
}
static gboolean handle_pipeline_message(GstBus *bus, GstMessage *msg, gpointer video_decoder)
{
SpiceGstDecoder *decoder = video_decoder;
switch(GST_MESSAGE_TYPE(msg)) {
case GST_MESSAGE_ERROR: {
GError *err = NULL;
gchar *debug_info = NULL;
gst_message_parse_error(msg, &err, &debug_info);
spice_warning("GStreamer error from element %s: %s",
GST_OBJECT_NAME(msg->src), err->message);
if (debug_info) {
SPICE_DEBUG("debug information: %s", debug_info);
g_free(debug_info);
}
g_clear_error(&err);
/* We won't be able to process any more frame anyway */
free_pipeline(decoder);
break;
}
case GST_MESSAGE_STREAM_START: {
gchar *filename = g_strdup_printf("spice-gtk-gst-pipeline-debug-%" G_GUINT32_FORMAT "-%s",
decoder->base.stream->id,
gst_opts[decoder->base.codec_type].name);
GST_DEBUG_BIN_TO_DOT_FILE(GST_BIN(decoder->pipeline),
GST_DEBUG_GRAPH_SHOW_ALL
| GST_DEBUG_GRAPH_SHOW_FULL_PARAMS
| GST_DEBUG_GRAPH_SHOW_STATES,
filename);
g_free(filename);
break;
}
case GST_MESSAGE_QOS: {
// seems sometimes it drops but do not emit qos msg, basically drops should be update the queue.
GstFormat format;
guint64 processed;
guint64 dropped;
gst_message_parse_qos_stats(msg, &format, &processed, &dropped);
printf("Feedback QOS MSG(%d): processed: %lu dropped: %lu \n", (int)format, processed, dropped);
break;
}
default:
/* not being handled */
break;
}
return TRUE;
}
static void app_source_setup(GstElement *pipeline G_GNUC_UNUSED,
GstElement *source,
SpiceGstDecoder *decoder)
{
GstCaps *caps;
/* - We schedule the frame display ourselves so set sync=false on appsink
* so the pipeline decodes them as fast as possible. This will also
* minimize the risk of frames getting lost when we rebuild the
* pipeline.
* - Set max-bytes=0 on appsrc so it does not drop frames that may be
* needed by those that follow.
*/
caps = gst_caps_from_string(gst_opts[decoder->base.codec_type].dec_caps);
g_object_set(source,
"caps", caps,
"is-live", LIVE, //add latency estimation
"format", GST_FORMAT_TIME,
"max-bytes", G_GINT64_CONSTANT(0),
"block", TRUE,
NULL);
gst_caps_unref(caps);
decoder->appsrc = GST_APP_SRC(gst_object_ref(source));
}
static GstPadProbeReturn event_probe(GstPad *pad,
GstPadProbeInfo *info, gpointer data)
{
SpiceGstDecoder *decoder = (SpiceGstDecoder*)data;
static GstClockTime last;
if (info->type & GST_PAD_PROBE_TYPE_BUFFER) { // Buffer arrived
GstBuffer *obj = GST_PAD_PROBE_INFO_BUFFER(info);
GstClockTime cur = gst_clock_get_time(decoder->clock);
gdouble rate = gst_guint64_to_gdouble(cur - last) / gst_guint64_to_gdouble(GST_BUFFER_DURATION(obj)); // rate is the ratio between actual procssing time to target rate
decoder->queue--;
if (GST_CLOCK_TIME_IS_VALID(last) && last != 0) {
if (decoder->avg_rate < 0.0) {
decoder->avg_rate = rate;
} else {
if (rate > 1.0) {
decoder->avg_rate = UPDATE_RUNNING_AVG/*_N*/ (decoder->avg_rate, rate);
} else {
decoder->avg_rate = UPDATE_RUNNING_AVG/*_P*/ (decoder->avg_rate, rate);
}
}
}
printf("BUFFERS QOS: queue: %u, AVG RATE: %f\n",decoder->queue, decoder->avg_rate);
last = cur;
} else { // qos & latency events
GstEvent *event = GST_PAD_PROBE_INFO_EVENT(info);
g_assert (GST_IS_EVENT(event));
if (GST_EVENT_TYPE(event) == GST_EVENT_QOS) { // QOS event as calculated by gstreamer, may cause behaviour changing in elements (i.e. dropping frames)
GstQOSType type;
gdouble proportion;
GstClockTimeDiff diff;
GstClockTime timestamp;
gst_event_parse_qos (event, &type, &proportion, &diff, ×tamp);
printf("GST_EVENT_QOS type %u, proportion %lf, diff %"
G_GINT64_FORMAT ", timestamp %" GST_TIME_FORMAT "\n", type,
proportion,
diff,
GST_TIME_ARGS (timestamp));
}
if (GST_EVENT_TYPE(event) == GST_EVENT_LATENCY) { // Tells sink to adjust their synchronisation with latency
GstClockTime latency;
gst_event_parse_latency (event, &latency);
printf("GST_EVENT_LATENCY: %lums\n", latency/1000/1000);
}
}
return GST_PAD_PROBE_OK;
}
static inline const char *gst_element_name(GstElement *element)
{
GstElementFactory *f = gst_element_get_factory(element);
return f ? GST_OBJECT_NAME(f) : GST_OBJECT_NAME(element);
}
// This function is used to set properties in dynamically added sink (if overlay is used), and setting a probe on the sink
static void
add_elem_cb(GstBin * pipeline, GstBin * bin, GstElement * element, SpiceGstDecoder *decoder)
{
SPICE_DEBUG("A new element was added to Gstreamer's pipeline (%s)",
gst_element_name(element));
char *name = gst_element_get_name(element);
spice_debug("Adding element: %s", name);
if (GST_IS_BASE_SINK(element)) {// && GST_OBJECT_FLAG_IS_SET(element, GST_ELEMENT_FLAG_SINK)
GstPad *pad;
pad = gst_element_get_static_pad(element, "sink");
gst_pad_add_probe(pad, GST_PAD_PROBE_TYPE_EVENT_UPSTREAM | GST_PAD_PROBE_TYPE_BUFFER, event_probe, decoder, NULL); //also buffers
g_object_set(element,
"sync", SYNC,
"qos", QOS,
"drop", DROP,
NULL);
gst_object_unref(pad);
spice_debug("^^^^SINK^^^^");
} else {
//just trying to set this in other elements
/*g_object_set(element,
"max-size-buffers", 0,
NULL);*/
}
g_free(name);
}
static gboolean create_pipeline(SpiceGstDecoder *decoder)
{
GstBus *bus;
GstElement *playbin, *sink;
SpiceGstPlayFlags flags;
GstCaps *caps;
playbin = gst_element_factory_make("playbin", "playbin");
if (playbin == NULL) {
spice_warning("error upon creation of 'playbin' element");
return FALSE;
}
/* Passing the pipeline to widget, try to get window handle and
* set the GstVideoOverlay interface, setting overlay to the window
* will happen only when prepare-window-handle message is received
*/
if (!hand_pipeline_to_widget(decoder->base.stream, GST_PIPELINE(playbin))) {
sink = gst_element_factory_make("appsink", "sink");
if (sink == NULL) {
spice_warning("error upon creation of 'appsink' element");
gst_object_unref(playbin);
return FALSE;
}
caps = gst_caps_from_string("video/x-raw,format=BGRx");
g_object_set(sink,
"caps", caps,
"sync", SYNC,
"drop", DROP,
NULL);
gst_caps_unref(caps);
g_object_set(playbin,
"video-sink", gst_object_ref(sink),
NULL);
decoder->appsink = GST_APP_SINK(sink);
} else {
/* handle has received, it means playbin will render directly into
* widget using the gstvideooverlay interface instead of app-sink.
*/
SPICE_DEBUG("Video is presented using gstreamer's GstVideoOverlay interface");
#if !GST_CHECK_VERSION(1,14,0)
/* Avoid using vaapisink if exist since vaapisink could be
* buggy when it is combined with playbin. changing its rank to
* none will make playbin to avoid of using it.
*/
GstRegistry *registry = NULL;
GstPluginFeature *vaapisink = NULL;
registry = gst_registry_get();
if (registry) {
vaapisink = gst_registry_lookup_feature(registry, "vaapisink");
}
if (vaapisink) {
gst_plugin_feature_set_rank(vaapisink, GST_RANK_NONE);
gst_object_unref(vaapisink);
}
#endif
//disable vaapi decoding
/*if (registry) {
vaapisink = gst_registry_lookup_feature(registry, "vaapih264dec");
}
if (vaapisink) {
gst_plugin_feature_set_rank(vaapisink, GST_RANK_NONE);
gst_object_unref(vaapisink);
}
if (registry) {
vaapisink = gst_registry_lookup_feature(registry, "vaapidecodebin");
}
if (vaapisink) {
gst_plugin_feature_set_rank(vaapisink, GST_RANK_NONE);
gst_object_unref(vaapisink);
}*/
// force sink
/*sink = gst_element_factory_make("xvimagesink", "sink");
g_object_set(playbin,
"video-sink", gst_object_ref(sink),
NULL);*/
}
g_signal_connect(playbin, "deep-element-added", G_CALLBACK(add_elem_cb), decoder);
g_signal_connect(playbin, "source-setup", G_CALLBACK(app_source_setup), decoder);
g_object_set(playbin,
"uri", "appsrc://",
NULL);
/* Disable audio in playbin */
g_object_get(playbin, "flags", &flags, NULL);
flags &= ~(GST_PLAY_FLAG_AUDIO | GST_PLAY_FLAG_TEXT);
g_object_set(playbin, "flags", flags, NULL);
g_warn_if_fail(decoder->appsrc == NULL);
decoder->pipeline = playbin;
if (decoder->appsink) {
GstAppSinkCallbacks appsink_cbs = { NULL };
appsink_cbs.new_sample = new_sample;
gst_app_sink_set_callbacks(decoder->appsink, &appsink_cbs, decoder, NULL);
gst_app_sink_set_max_buffers(decoder->appsink, MAX_DECODED_FRAMES);
gst_app_sink_set_drop(decoder->appsink, FALSE);
}
bus = gst_pipeline_get_bus(GST_PIPELINE(decoder->pipeline));
gst_bus_add_watch(bus, handle_pipeline_message, decoder);
gst_object_unref(bus);
decoder->clock = gst_pipeline_get_clock(GST_PIPELINE(decoder->pipeline));
if (gst_element_set_state(decoder->pipeline, GST_STATE_PLAYING) == GST_STATE_CHANGE_FAILURE) {
SPICE_DEBUG("GStreamer error: Unable to set the pipeline to the playing state.");
free_pipeline(decoder);
return FALSE;
}
return TRUE;
}
/* ---------- VideoDecoder's public API ---------- */
static void spice_gst_decoder_reschedule(VideoDecoder *video_decoder)
{
SpiceGstDecoder *decoder = (SpiceGstDecoder*)video_decoder;
if (!decoder->appsink) {
return;
}
guint timer_id;
g_mutex_lock(&decoder->queues_mutex);
timer_id = decoder->timer_id;
decoder->timer_id = 0;
g_mutex_unlock(&decoder->queues_mutex);
if (timer_id != 0) {
g_source_remove(timer_id);
}
schedule_frame(decoder);
}
/* main context */
static void spice_gst_decoder_destroy(VideoDecoder *video_decoder)
{
SpiceGstDecoder *decoder = (SpiceGstDecoder*)video_decoder;
/* Stop and free the pipeline to ensure there will not be any further
* new_sample() call (clearing thread-safety concerns).
*/
free_pipeline(decoder);
/* Even if we kept the decoder around, once we return the stream will be
* destroyed making it impossible to display frames. So cancel any
* scheduled display_frame() call and drop the queued frames.
*/
if (decoder->timer_id) {
g_source_remove(decoder->timer_id);
}
g_mutex_clear(&decoder->queues_mutex);
g_queue_free_full(decoder->decoding_queue, (GDestroyNotify)free_gst_frame);
if (decoder->display_frame) {
free_gst_frame(decoder->display_frame);
}
g_free(decoder);
/* Don't call gst_deinit() as other parts of the client
* may still be using GStreamer.
*/
}
/* spice_gst_decoder_queue_frame() queues the SpiceFrame for decoding and
* displaying. The steps it goes through are as follows:
*
* 1) frame->data, which contains the compressed frame data, is wrapped in a GstBuffer
* (encoded_buffer) which owns the SpiceFrame.
* 2) A SpiceGstFrame is created to keep track of SpiceFrame (encoded_frame),
* and additional metadata among which GStreamer's encoded_buffer the
* refcount of which is incremented. The SpiceGstFrame is then pushed into
* the decoding_queue.
*
* If GstVideoOverlay is used (window handle was obtained successfully at the widget):
* 3) Decompressed frames will be rendered to widget directly from GStreamer's pipeline
* using some GStreamer sink plugin which implements the GstVideoOverlay interface
* (last step).
* 4) As soon as GStreamer's pipeline no longer needs the compressed frame it will
* unref the encoded_buffer.
* 5) Once a decoded buffer arrives to the sink sink_event_probe() will pop
* its matching SpiceGstFrame from the decoding_queue and free it using
* free_gst_frame(). This will also unref the encoded_buffer which will
* allow GStreamer to call spice_frame_free() and free its encoded_frame.
*
* Otherwise appsink is used:
* 3) Once the decompressed frame is available the GStreamer pipeline calls
* new_sample() in the GStreamer thread.
* 4) new_sample() then increments the pending_samples count and calls
* schedule_frame().
* 5) schedule_frame() is called whenever a new frame might need to be
* displayed. If that is the case and pending_samples is non-zero it calls
* fetch_pending_sample().
* 6) fetch_pending_sample() grabs GStreamer's latest sample and then calls
* get_decoded_frame() which compares the GStreamer's buffer timestamp to
* gstframe->encoded_frame->mm_time to match it with a decoding_queue
* entry.
* 7) fetch_pending_sample() then attaches the sample to the SpiceGstFrame,
* and sets display_frame.
* 8) schedule_frame() then uses display_frame->encoded_frame->mm_time to
* arrange for display_frame() to be called, in the main thread, at the
* right time.
* 9) display_frame() uses SpiceGstFrame from display_frame and calls
* stream_display_frame().
* 10) display_frame() then calls free_gst_frame() to free the SpiceGstFrame
* and unref the encoded_buffer which allows GStreamer to call
* spice_frame_free() and free its encoded_frame.
*/
static gboolean spice_gst_decoder_queue_frame(VideoDecoder *video_decoder,
SpiceFrame *frame, int margin)
{
SpiceGstDecoder *decoder = (SpiceGstDecoder*)video_decoder;
if (frame->size == 0) {
SPICE_DEBUG("got an empty frame buffer!");
spice_frame_free(frame);
return TRUE;
}
if (spice_mmtime_diff(frame->mm_time, decoder->last_mm_time) < 0) {
SPICE_DEBUG("new-frame-time < last-frame-time (%u < %u):"
" resetting stream",
frame->mm_time, decoder->last_mm_time);
/* Let GStreamer deal with the frame anyway */
}
decoder->last_mm_time = frame->mm_time;
if (margin < 0 &&
decoder->base.codec_type == SPICE_VIDEO_CODEC_TYPE_MJPEG) {
/* Dropping MJPEG frames has no impact on those that follow and
* saves CPU so do it.
*/
SPICE_DEBUG("dropping a late MJPEG frame");
spice_frame_free(frame);
return TRUE;
}
if (decoder->pipeline == NULL) {
/* An error occurred, causing the GStreamer pipeline to be freed */
spice_warning("An error occurred, stopping the video stream");
spice_frame_free(frame);
return FALSE;
}
if (decoder->appsrc == NULL) {
spice_warning("Error: Playbin has not yet initialized the Appsrc element");
stream_dropped_frame_on_playback(decoder->base.stream);
spice_frame_free(frame);
return TRUE;
}
/* frame ownership is moved to the buffer */
GstBuffer *buffer = gst_buffer_new_wrapped_full(GST_MEMORY_FLAG_PHYSICALLY_CONTIGUOUS,
frame->data, frame->size, 0, frame->size,
frame, (GDestroyNotify) spice_frame_free);
// GstClockTime pts = gst_clock_get_time(decoder->clock) - gst_element_get_base_time(decoder->pipeline) + ((uint64_t)MAX(0, margin)) * 1000 * 1000;
GstClockTime pts = gst_clock_get_time(decoder->clock) - gst_element_get_base_time(decoder->pipeline); //ignore margin and audio sync, this is actually kind of arrival time based
//GstClockTime pts = GST_CLOCK_TIME_NONE;
GST_BUFFER_DURATION(buffer) = GST_CLOCK_TIME_NONE;
GST_BUFFER_DTS(buffer) = GST_CLOCK_TIME_NONE;
GST_BUFFER_PTS(buffer) = pts;
#if GST_CHECK_VERSION(1,14,0)
gst_buffer_add_reference_timestamp_meta(buffer, gst_static_caps_get(&stream_reference),
pts, GST_CLOCK_TIME_NONE);
#endif
SpiceGstFrame *gst_frame = create_gst_frame(buffer, frame);
g_mutex_lock(&decoder->queues_mutex);
gst_frame->queue_len = decoder->decoding_queue->length;
g_queue_push_tail(decoder->decoding_queue, gst_frame);
g_mutex_unlock(&decoder->queues_mutex);
if (gst_app_src_push_buffer(decoder->appsrc, buffer) != GST_FLOW_OK) {
SPICE_DEBUG("GStreamer error: unable to push frame");
stream_dropped_frame_on_playback(decoder->base.stream);
} else {
decoder->queue++;
}
return TRUE;
}
static gboolean gstvideo_init(void)
{
static int success = 0;
if (!success) {
GError *err = NULL;
if (gst_init_check(NULL, NULL, &err)) {
success = 1;
} else {
spice_warning("Disabling GStreamer video support: %s", err->message);
g_clear_error(&err);
success = -1;
}
}
return success > 0;
}
G_GNUC_INTERNAL
VideoDecoder* create_gstreamer_decoder(int codec_type, display_stream *stream)
{
SpiceGstDecoder *decoder = NULL;
g_return_val_if_fail(VALID_VIDEO_CODEC_TYPE(codec_type), NULL);
if (gstvideo_init()) {
decoder = g_new0(SpiceGstDecoder, 1);
decoder->base.destroy = spice_gst_decoder_destroy;
decoder->base.reschedule = spice_gst_decoder_reschedule;
decoder->base.queue_frame = spice_gst_decoder_queue_frame;
decoder->base.codec_type = codec_type;
decoder->base.stream = stream;
decoder->last_mm_time = stream_get_time(stream);
g_mutex_init(&decoder->queues_mutex);
decoder->decoding_queue = g_queue_new();
decoder->avg_rate = -1;
if (!create_pipeline(decoder)) {
decoder->base.destroy((VideoDecoder*)decoder);
decoder = NULL;
}
}
return (VideoDecoder*)decoder;
}
static void gstvideo_debug_available_decoders(int codec_type,
GList *all_decoders,
GList *codec_decoders)
{
GList *l;
GString *msg = g_string_new(NULL);
/* Print list of available decoders to make debugging easier */
g_string_printf(msg, "From %3u video decoder elements, %2u can handle caps %12s: ",
g_list_length(all_decoders), g_list_length(codec_decoders),
gst_opts[codec_type].dec_caps);
for (l = codec_decoders; l != NULL; l = l->next) {
GstPluginFeature *pfeat = GST_PLUGIN_FEATURE(l->data);
g_string_append_printf(msg, "%s, ", gst_plugin_feature_get_name(pfeat));
}
/* Drop trailing ", " */
g_string_truncate(msg, msg->len - 2);
spice_debug("%s", msg->str);
g_string_free(msg, TRUE);
}
G_GNUC_INTERNAL
gboolean gstvideo_has_codec(int codec_type)
{
GList *all_decoders, *codec_decoders;
GstCaps *caps;
GstElementFactoryListType type;
g_return_val_if_fail(gstvideo_init(), FALSE);
g_return_val_if_fail(VALID_VIDEO_CODEC_TYPE(codec_type), FALSE);
type = GST_ELEMENT_FACTORY_TYPE_DECODER |
GST_ELEMENT_FACTORY_TYPE_MEDIA_VIDEO |
GST_ELEMENT_FACTORY_TYPE_MEDIA_IMAGE;
all_decoders = gst_element_factory_list_get_elements(type, GST_RANK_NONE);
if (all_decoders == NULL) {
spice_debug("No video decoders from GStreamer for %s were found",
gst_opts[codec_type].name);
return FALSE;
}
caps = gst_caps_from_string(gst_opts[codec_type].dec_caps);
codec_decoders = gst_element_factory_list_filter(all_decoders, caps, GST_PAD_SINK, FALSE);
gst_caps_unref(caps);
if (codec_decoders == NULL) {
spice_debug("From %u decoders, none can handle '%s'",
g_list_length(all_decoders), gst_opts[codec_type].dec_caps);
gst_plugin_feature_list_free(all_decoders);
return FALSE;
}
if (spice_util_get_debug())
gstvideo_debug_available_decoders(codec_type, all_decoders, codec_decoders);
gst_plugin_feature_list_free(codec_decoders);
gst_plugin_feature_list_free(all_decoders);
return TRUE;
}
|