+RESULT eServiceMP3::seekRelative(int direction, pts_t to)
+{
+ if (!m_gst_pipeline)
+ return -1;
+
+ pts_t ppos;
+ getPlayPosition(ppos);
+ ppos += to * direction;
+ if (ppos < 0)
+ ppos = 0;
+ seekTo(ppos);
+
+ return 0;
+}
+
+RESULT eServiceMP3::getPlayPosition(pts_t &pts)
+{
+ if (!m_gst_pipeline)
+ return -1;
+ if (m_state != stRunning)
+ return -1;
+
+ GstFormat fmt = GST_FORMAT_TIME;
+ gint64 len;
+
+ if (!gst_element_query_position(m_gst_pipeline, &fmt, &len))
+ return -1;
+
+ /* len is in nanoseconds. we have 90 000 pts per second. */
+ pts = len / 11111;
+ return 0;
+}
+
+RESULT eServiceMP3::setTrickmode(int trick)
+{
+ /* trickmode is not yet supported by our dvbmediasinks. */
+ return -1;
+}
+
+RESULT eServiceMP3::isCurrentlySeekable()
+{
+ return 1;
+}
+
+RESULT eServiceMP3::info(ePtr<iServiceInformation>&i)
+{
+ i = this;
+ return 0;
+}
+
+RESULT eServiceMP3::getName(std::string &name)
+{
+ name = m_filename;
+ size_t n = name.rfind('/');
+ if (n != std::string::npos)
+ name = name.substr(n + 1);
+ return 0;
+}
+
+int eServiceMP3::getInfo(int w)
+{
+ gchar *tag = 0;
+
+ switch (w)
+ {
+ case sTitle:
+ case sArtist:
+ case sAlbum:
+ case sComment:
+ case sTracknumber:
+ case sGenre:
+ case sVideoType:
+ return resIsString;
+ case sCurrentTitle:
+ tag = GST_TAG_TRACK_NUMBER;
+ break;
+ case sTotalTitles:
+ tag = GST_TAG_TRACK_COUNT;
+ break;
+ default:
+ return resNA;
+ }
+
+ if (!m_stream_tags || !tag)
+ return 0;
+
+ guint value;
+ if (gst_tag_list_get_uint(m_stream_tags, tag, &value))
+ return (int) value;
+
+ return 0;
+
+}
+
+std::string eServiceMP3::getInfoString(int w)
+{
+ gchar *tag = 0;
+ switch (w)
+ {
+ case sTitle:
+ tag = GST_TAG_TITLE;
+ break;
+ case sArtist:
+ tag = GST_TAG_ARTIST;
+ break;
+ case sAlbum:
+ tag = GST_TAG_ALBUM;
+ break;
+ case sComment:
+ tag = GST_TAG_COMMENT;
+ break;
+ case sTracknumber:
+ tag = GST_TAG_TRACK_NUMBER;
+ break;
+ case sGenre:
+ tag = GST_TAG_GENRE;
+ break;
+ case sVideoType:
+ tag = GST_TAG_VIDEO_CODEC;
+ break;
+ default:
+ return "";
+ }
+
+ if (!m_stream_tags || !tag)
+ return "";
+
+ gchar *value;
+
+ if (gst_tag_list_get_string(m_stream_tags, tag, &value))
+ {
+ std::string res = value;
+ g_free(value);
+ return res;
+ }
+
+ return "";
+}
+
+RESULT eServiceMP3::audioChannel(ePtr<iAudioChannelSelection> &ptr)
+{
+ ptr = this;
+ return 0;
+}
+
+RESULT eServiceMP3::audioTracks(ePtr<iAudioTrackSelection> &ptr)
+{
+ ptr = this;
+ return 0;
+}
+
+RESULT eServiceMP3::subtitle(ePtr<iSubtitleOutput> &ptr)
+{
+ ptr = this;
+ return 0;
+}
+
+int eServiceMP3::getNumberOfTracks()
+{
+ return m_audioStreams.size();
+}
+
+int eServiceMP3::getCurrentTrack()
+{
+ return m_currentAudioStream;
+}
+
+RESULT eServiceMP3::selectTrack(unsigned int i)
+{
+ int ret = selectAudioStream(i);
+ /* flush */
+ pts_t ppos;
+ getPlayPosition(ppos);
+ seekTo(ppos);
+
+ return ret;
+}
+
+int eServiceMP3::selectAudioStream(int i)
+{
+ gint nb_sources;
+ GstPad *active_pad;
+ GstElement *selector = gst_bin_get_by_name(GST_BIN(m_gst_pipeline),"switch_audio");
+ if ( !selector)
+ {
+ eDebug("can't switch audio tracks! gst-plugin-selector needed");
+ return -1;
+ }
+ g_object_get (G_OBJECT (selector), "n-pads", &nb_sources, NULL);
+ if ( i >= m_audioStreams.size() || i >= nb_sources || m_currentAudioStream >= m_audioStreams.size() )
+ return -2;
+ char sinkpad[8];
+ sprintf(sinkpad, "sink%d", i);
+ g_object_set (G_OBJECT (selector), "active-pad", gst_element_get_pad (selector, sinkpad), NULL);
+ g_object_get (G_OBJECT (selector), "active-pad", &active_pad, NULL);
+ gchar *name;
+ name = gst_pad_get_name (active_pad);
+ eDebug ("switched audio to (%s)", name);
+ g_free(name);
+ m_currentAudioStream = i;
+ return 0;
+}
+
+int eServiceMP3::getCurrentChannel()
+{
+ return STEREO;
+}
+
+RESULT eServiceMP3::selectChannel(int i)
+{
+ eDebug("eServiceMP3::selectChannel(%i)",i);
+ return 0;
+}
+
+RESULT eServiceMP3::getTrackInfo(struct iAudioTrackInfo &info, unsigned int i)
+{
+// eDebug("eServiceMP3::getTrackInfo(&info, %i)",i);
+ if (i >= m_audioStreams.size())
+ return -2;
+ if (m_audioStreams[i].type == audioStream::atMP2)
+ info.m_description = "MP2";
+ else if (m_audioStreams[i].type == audioStream::atMP3)
+ info.m_description = "MP3";
+ else if (m_audioStreams[i].type == audioStream::atAC3)
+ info.m_description = "AC3";
+ else if (m_audioStreams[i].type == audioStream::atAAC)
+ info.m_description = "AAC";
+ else if (m_audioStreams[i].type == audioStream::atDTS)
+ info.m_description = "DTS";
+ else if (m_audioStreams[i].type == audioStream::atPCM)
+ info.m_description = "PCM";
+ else if (m_audioStreams[i].type == audioStream::atOGG)
+ info.m_description = "OGG";
+ else
+ info.m_description = "???";
+ if (info.m_language.empty())
+ info.m_language = m_audioStreams[i].language_code;
+ return 0;
+}
+
+void eServiceMP3::gstBusCall(GstBus *bus, GstMessage *msg)
+{
+ if (!msg)
+ return;
+ gchar *sourceName;
+ GstObject *source;
+
+ source = GST_MESSAGE_SRC(msg);
+ sourceName = gst_object_get_name(source);
+
+ if (gst_message_get_structure(msg))
+ {
+ gchar *string = gst_structure_to_string(gst_message_get_structure(msg));
+ eDebug("gst_message from %s: %s", sourceName, string);
+ g_free(string);
+ }
+ else
+ eDebug("gst_message from %s: %s (without structure)", sourceName, GST_MESSAGE_TYPE_NAME(msg));
+
+ switch (GST_MESSAGE_TYPE (msg))
+ {
+ case GST_MESSAGE_EOS:
+ m_event((iPlayableService*)this, evEOF);
+ break;
+ case GST_MESSAGE_ERROR:
+ {
+ gchar *debug;
+ GError *err;
+
+ gst_message_parse_error (msg, &err, &debug);
+ g_free (debug);
+ eWarning("Gstreamer error: %s (%i)", err->message, err->code );
+ if ( err->domain == GST_STREAM_ERROR && err->code == GST_STREAM_ERROR_DECODE )
+ {
+ if ( g_strrstr(sourceName, "videosink") )
+ m_event((iPlayableService*)this, evUser+11);
+ }
+ g_error_free(err);
+ /* TODO: signal error condition to user */
+ break;
+ }
+ case GST_MESSAGE_TAG:
+ {
+ GstTagList *tags, *result;
+ gst_message_parse_tag(msg, &tags);
+
+ result = gst_tag_list_merge(m_stream_tags, tags, GST_TAG_MERGE_PREPEND);
+ if (result)
+ {
+ if (m_stream_tags)
+ gst_tag_list_free(m_stream_tags);
+ m_stream_tags = result;
+ }
+ gchar *g_audiocodec;
+ if (gst_tag_list_get_string(tags, GST_TAG_AUDIO_CODEC, &g_audiocodec) && m_audioStreams.size())
+ {
+ std::vector<audioStream>::iterator IterAudioStream = m_audioStreams.begin();
+ while ( IterAudioStream->language_code.length() && IterAudioStream != m_audioStreams.end())
+ IterAudioStream++;
+ if ( g_strrstr(g_audiocodec, "MPEG-1 layer 2") )
+ IterAudioStream->type = audioStream::atMP2;
+ else if ( g_strrstr(g_audiocodec, "MPEG-1 layer 3") )
+ IterAudioStream->type = audioStream::atMP3;
+ else if ( g_strrstr(g_audiocodec, "AC-3 audio") )
+ IterAudioStream->type = audioStream::atAC3;
+ else if ( g_strrstr(g_audiocodec, "Uncompressed 16-bit PCM audio") )
+ IterAudioStream->type = audioStream::atPCM;
+ gchar *g_language;
+ if ( gst_tag_list_get_string(tags, GST_TAG_LANGUAGE_CODE, &g_language) )
+ IterAudioStream->language_code = std::string(g_language);
+ g_free (g_language);
+ g_free (g_audiocodec);
+ }
+ break;
+ }
+ default:
+ break;
+ }
+ g_free (sourceName);
+}
+
+GstBusSyncReply eServiceMP3::gstBusSyncHandler(GstBus *bus, GstMessage *message, gpointer user_data)
+{
+ eServiceMP3 *_this = (eServiceMP3*)user_data;
+ _this->m_pump.send(1);
+ /* wake */
+ return GST_BUS_PASS;
+}
+
+void eServiceMP3::gstCBpadAdded(GstElement *decodebin, GstPad *pad, gpointer user_data)
+{
+ eServiceMP3 *_this = (eServiceMP3*)user_data;
+ GstBin *pipeline = GST_BIN(_this->m_gst_pipeline);
+ gchar *name;
+ name = gst_pad_get_name (pad);
+ eDebug ("A new pad %s was created", name);
+ if (g_strrstr(name,"audio")) // mpegdemux, matroskademux, avidemux use video_nn with n=0,1,.., flupsdemux uses stream id
+ {
+ GstElement *selector = gst_bin_get_by_name(pipeline , "switch_audio" );
+ audioStream audio;
+ audio.pad = pad;
+ _this->m_audioStreams.push_back(audio);
+ if ( selector )
+ {
+ gst_pad_link(pad, gst_element_get_request_pad (selector, "sink%d"));
+ if ( _this->m_audioStreams.size() == 1 )
+ {
+ _this->selectAudioStream(0);
+ gst_element_set_state (_this->m_gst_pipeline, GST_STATE_PLAYING);
+ }
+ else
+ g_object_set (G_OBJECT (selector), "select-all", FALSE, NULL);
+ }
+ else
+ gst_pad_link(pad, gst_element_get_static_pad(gst_bin_get_by_name(pipeline,"queue_audio"), "sink"));
+ }
+ if (g_strrstr(name,"video"))
+ {
+ gst_pad_link(pad, gst_element_get_static_pad(gst_bin_get_by_name(pipeline,"queue_video"), "sink"));
+ }
+ if (g_strrstr(name,"subtitle"))
+ {
+// GstCaps *caps;
+// const GstStructure *structure;
+// caps = gst_pad_get_caps(name);
+// structure = gst_caps_get_structure(caps, 0);
+ char elemname[17];
+ sprintf(elemname, "%s_pars", name);
+ GstElement *parser = gst_element_factory_make("ssaparse", elemname);
+ eDebug ("ssaparse %s = %p", elemname, parser);
+ sprintf(elemname, "%s_sink", name);
+ GstElement *sink = gst_element_factory_make("fakesink", elemname);
+ eDebug ("fakesink %s = %p", elemname, sink);
+ g_object_set (G_OBJECT(sink), "signal-handoffs", TRUE, NULL);
+ gst_bin_add_many(pipeline, parser, sink, NULL);
+ GstPadLinkReturn res = gst_pad_link(pad, gst_element_get_static_pad(parser, "sink"));
+ eDebug ("parser link = %d", res);
+ res = gst_element_link(parser, sink);
+ eDebug ("sink link = %d", res);
+ g_signal_connect(sink, "handoff", G_CALLBACK(gstCBsubtitleAvail), _this);
+ subtitleStream subs;
+ subs.element = sink;
+ _this->m_subtitleStreams.push_back(subs);
+ }
+ g_free (name);
+}
+
+void eServiceMP3::gstCBfilterPadAdded(GstElement *filter, GstPad *pad, gpointer user_data)
+{
+ eServiceMP3 *_this = (eServiceMP3*)user_data;
+ GstElement *decoder = gst_bin_get_by_name(GST_BIN(_this->m_gst_pipeline),"decoder");
+ gst_pad_link(pad, gst_element_get_static_pad (decoder, "sink"));
+}
+
+void eServiceMP3::gstCBnewPad(GstElement *decodebin, GstPad *pad, gboolean last, gpointer user_data)
+{
+ eServiceMP3 *_this = (eServiceMP3*)user_data;
+ GstCaps *caps;
+ GstStructure *str;
+ GstPad *audiopad;
+
+ /* only link once */
+ GstElement *audio = gst_bin_get_by_name(GST_BIN(_this->m_gst_pipeline),"audiobin");
+ audiopad = gst_element_get_static_pad (audio, "sink");
+ if ( !audiopad || GST_PAD_IS_LINKED (audiopad)) {
+ eDebug("audio already linked!");
+ g_object_unref (audiopad);
+ return;
+ }
+
+ /* check media type */
+ caps = gst_pad_get_caps (pad);
+ str = gst_caps_get_structure (caps, 0);
+ eDebug("gst new pad! %s", gst_structure_get_name (str));
+
+ if (!g_strrstr (gst_structure_get_name (str), "audio")) {
+ gst_caps_unref (caps);
+ gst_object_unref (audiopad);
+ return;
+ }
+
+ gst_caps_unref (caps);
+ gst_pad_link (pad, audiopad);
+}
+
+void eServiceMP3::gstCBunknownType(GstElement *decodebin, GstPad *pad, GstCaps *caps, gpointer user_data)
+{
+ GstStructure *str;
+
+ /* check media type */
+ caps = gst_pad_get_caps (pad);
+ str = gst_caps_get_structure (caps, 0);
+ eDebug("unknown type: %s - this can't be decoded.", gst_structure_get_name (str));
+ gst_caps_unref (caps);
+}
+
+void eServiceMP3::gstPoll(const int&)
+{
+ /* ok, we have a serious problem here. gstBusSyncHandler sends
+ us the wakup signal, but likely before it was posted.
+ the usleep, an EVIL HACK (DON'T DO THAT!!!) works around this.
+
+ I need to understand the API a bit more to make this work
+ proplerly. */
+ usleep(1);
+
+ GstBus *bus = gst_pipeline_get_bus (GST_PIPELINE (m_gst_pipeline));
+ GstMessage *message;
+ while ((message = gst_bus_pop (bus)))
+ {
+ gstBusCall(bus, message);
+ gst_message_unref (message);
+ }
+}
+
+eAutoInitPtr<eServiceFactoryMP3> init_eServiceFactoryMP3(eAutoInitNumbers::service+1, "eServiceFactoryMP3");
+#else
+#warning gstreamer not available, not building media player
+#endif
+
+void eServiceMP3::gstCBsubtitleAvail(GstElement *element, GstBuffer *buffer, GstPad *pad, gpointer user_data)
+{
+ const char *text = (unsigned char *)GST_BUFFER_DATA(buffer);
+ eServiceMP3 *_this = (eServiceMP3*)user_data;
+ gchar *sourceName;
+ sourceName = gst_object_get_name(GST_OBJECT(element));
+ if ( _this->m_subtitle_widget && _this->m_subtitleStreams.at(_this->m_currentSubtitleStream).element == element)
+ {
+ eDVBTeletextSubtitlePage page;
+ gRGB rgbcol(0xD0,0xD0,0xD0);
+ page.m_elements.push_back(eDVBTeletextSubtitlePageElement(rgbcol, text));
+ (_this->m_subtitle_widget)->setPage(page);
+ }
+ else
+ eDebug("on inactive element: %s (%p) saw subtitle: %s",sourceName, element, text);
+ return TRUE;
+}
+
+RESULT eServiceMP3::enableSubtitles(eWidget *parent, ePyObject tuple)
+{
+ eDebug("eServiceMP3::enableSubtitles");
+
+ ePyObject entry;
+ int tuplesize = PyTuple_Size(tuple);
+ int type = 0;
+ int page, magazine, pid;
+
+ if (!PyTuple_Check(tuple))
+ goto error_out;
+
+ if (tuplesize < 1)
+ goto error_out;
+
+ entry = PyTuple_GET_ITEM(tuple, 0);
+
+ if (!PyInt_Check(entry))
+ goto error_out;
+
+ type = PyInt_AsLong(entry);
+
+ entry = PyTuple_GET_ITEM(tuple, 1);
+ if (!PyInt_Check(entry))
+ goto error_out;
+ pid = PyInt_AsLong(entry);
+
+ m_subtitle_widget = new eSubtitleWidget(parent);
+ m_subtitle_widget->resize(parent->size()); /* full size */
+ m_currentSubtitleStream = pid;
+
+ return 0;
+error_out:
+ eDebug("enableSubtitles needs a tuple as 2nd argument!\n"
+ "for gst subtitles (2, subtitle_stream_count)"
+ return -1;
+}
+
+RESULT eServiceMP3::disableSubtitles(eWidget *parent)
+{
+ eDebug("eServiceMP3::disableSubtitles");
+ delete m_subtitle_widget;
+ m_subtitle_widget = 0;
+ return 0;
+}
+
+PyObject *eServiceMP3::getCachedSubtitle()
+{
+ eDebug("eServiceMP3::eDVBServicePlay");
+ Py_RETURN_NONE;
+}
+
+PyObject *eServiceMP3::getSubtitleList()
+{
+ eDebug("eServiceMP3::getSubtitleList");
+
+ ePyObject l = PyList_New(0);
+ gchar *sourceName;
+ int stream_count = 0;
+
+ for (std::vector<subtitleStream>::iterator IterSubtitleStream(m_subtitleStreams.begin()); IterSubtitleStream != m_subtitleStreams.end(); ++IterSubtitleStream)
+ {
+ ePyObject tuple = PyTuple_New(5);
+ PyTuple_SET_ITEM(tuple, 0, PyInt_FromLong(2));
+ PyTuple_SET_ITEM(tuple, 1, PyInt_FromLong(stream_count));
+ PyTuple_SET_ITEM(tuple, 2, PyInt_FromLong(0));
+ PyTuple_SET_ITEM(tuple, 3, PyInt_FromLong(0));
+ sourceName = gst_object_get_name(GST_OBJECT (IterSubtitleStream->element));
+ PyTuple_SET_ITEM(tuple, 4, PyString_FromString(sourceName));
+ PyList_Append(l, tuple);
+ Py_DECREF(tuple);
+ stream_count++;
+ }
+
+ return l;
+}