使用gstreamer-0.10动态添加到加法器时出现的问题
我正在为dub mp4视频构建一个应用程序。我的目标是当用户观看mp4视频时,用户可以录制自己的声音。然后,我的管道将这个音频文件添加到管道加法器元素中,以便同时侦听所有记录。之后,我的目标是将音频和视频混合到mp4文件中。 嗯。我的问题是当录制停止,我向加法器添加一个文件时,管道崩溃。我这样做: 1.暂停管道 2.-从管道中移除“记录箱” 3.-将状态NULL设置为“记录箱” 4.-将EOS事件发送到“记录箱” 5.-将“音频箱”(加法器)添加到管道中 6.-将所需的静音插入wav文件 7.-将音频文件添加到“音频箱”//的加法器中,问题可能出在这里 我的Vala/精灵代码如下:使用gstreamer-0.10动态添加到加法器时出现的问题,gstreamer,Gstreamer,我正在为dub mp4视频构建一个应用程序。我的目标是当用户观看mp4视频时,用户可以录制自己的声音。然后,我的管道将这个音频文件添加到管道加法器元素中,以便同时侦听所有记录。之后,我的目标是将音频和视频混合到mp4文件中。 嗯。我的问题是当录制停止,我向加法器添加一个文件时,管道崩溃。我这样做: 1.暂停管道 2.-从管道中移除“记录箱” 3.-将状态NULL设置为“记录箱” 4.-将EOS事件发送到“记录箱” 5.-将“音频箱”(加法器)添加到管道中 6.-将所需的静音插入wav文件 7.
uses
Gtk
Gst
Gee
class AudioFilesSrc : Gst.Bin
wavparse: Element
src:Element
audioconvert: Element
srcpad: Pad
sinkpad: Pad
def OnDynamicPad (element:Element, zz:Pad)
var opad = audioconvert.get_static_pad("sink");
zz.link(opad);
def open(p:Gst.Bin,mixer:Element,s1:string)
src = Gst.ElementFactory.make("filesrc", "src1");
wavparse = ElementFactory.make("wavparse","wavparse");
audioconvert = ElementFactory.make("audioconvert","audioconvert");
wavparse.pad_added.connect(OnDynamicPad);
this.add_many(src,wavparse,audioconvert);
src.link_many(wavparse,audioconvert);
src.set("location",s1);
// añade este bin a pipeline general
p.add(this)
//busca la salida de audioconvert1 y conviertela en la salida del bin
srcpad = new Gst.GhostPad("src", audioconvert.get_static_pad("src"));
this.add_pad(srcpad);
print ".-.----------------------------- abierto "+s1
def conecta(mixer:Element)
//crea una entrada al mixer
sinkpad = mixer.get_request_pad("sink%d")
//this.get_pad("src").link(this.sinkpad)
srcpad.link(this.sinkpad)
//this.srcpad.set_blocked(false)
print ".-.----------------------------- conectado"
def close(p:Gst.Bin,mixer:Element)
p.set_state(State.PAUSED)
this.srcpad.set_blocked(true)
this.set_state(State.NULL)
this.srcpad.unlink(this.sinkpad)
mixer.release_request_pad (this.sinkpad)
init
Gtk.init (ref args)
Gst.init (ref args);
var prueba = new ventana ()
prueba.show_all ()
Gtk.main ();
class ventana : Window
drawing_area:private DrawingArea;
videopipeline: private Pipeline ;
recordbin: Gst.Bin
videobin: Gst.Bin
audiobin: Gst.Bin
volume: private Element ;
videosrc :private Element;
videosink: private Element;
videodec: private Element;
vaudiosink: private Element;
vaudioparser: Element;
vaudiodec: Element;
vaudioadder: private Element;
vaudioarchivos:list of AudioFilesSrc
recordsrc :private Element;
recordsink: private Element;
recordconvert: private Element;
recordencoder: private Element;
comienzo_grabacion:float
xid :private ulong ;
reloj:uint
position: float
duracion:float
bus:Gst.Bus
bus2:Gst.Bus
msg:Gst.Message
msg2:Gst.Message
seek_enabled:bool
seek_enabled2:bool
scale_1:Scale
estado:string
lugar:int
numgrab:int=0
archivos_audio:list of string
button:Button
button1:Button
button2:Button
button3:Button
button4:Button
button5:Button
button6:Button
button7:Button
button8:Button
init
reloj = Timeout.add(1000, mover)
title = "Bikoizketa"
default_height = 250
default_width = 450
window_position = WindowPosition.CENTER
comienzo_grabacion=-1
// video pipeline
duracion=Gst.CLOCK_TIME_NONE;
this.videopipeline = new Pipeline ("mypipeline");
videobin= new Gst.Bin("videobin")
this.videosrc = ElementFactory.make ("filesrc", "filesrc2")
this.videosrc.set("location","gontzal3.mp4");
this.videodec = ElementFactory.make ("decodebin", "dec");
this.videosink = ElementFactory.make ("xvimagesink", "videosink");
this.videosink.set("force-aspect-ratio",true)
this.videodec.pad_added.connect(OnDynamicPad);
this.videobin.add_many (videosrc,videodec,videosink)
this.videosrc.link_many (videodec,videosink)
this.videopipeline.add(videobin)
this.audiobin= new Gst.Bin("audiobin")
this.vaudioadder = ElementFactory.make("adder","mixer");
this.volume= ElementFactory.make("volume","volume");
this.volume.set_property("volume",0.5)
this.vaudiosink= ElementFactory.make("alsasink","alsasink");
this.audiobin.add_many (vaudioadder,volume,vaudiosink)
vaudioadder.link_many(volume,vaudiosink)
this.videopipeline.add(audiobin)
//añadiendo archivos
vaudioarchivos=new list of AudioFilesSrc
vaudioarchivos.add (new AudioFilesSrc())
vaudioarchivos.last().open(audiobin,vaudioadder,"silencios.wav")
vaudioarchivos.last().conecta(vaudioadder)
numgrab++
this.recordbin= new Gst.Bin("recorder")
this.recordsrc= ElementFactory.make ("alsasrc","alsasrc")
this.recordconvert=ElementFactory.make ("audioconvert","audioconvert")
this.recordencoder = Gst.ElementFactory.make("wavenc", "encoder")
this.recordsink= ElementFactory.make ("filesink","filesink")
this.recordsink.set ("location","grabacionx.wav")
this.recordbin.add_many (this.recordsrc,this.recordconvert, this.recordencoder, this.recordsink);
this.recordsrc.link_many(this.recordconvert,recordencoder,recordsink)
//videopipeline.add(recordbin)
// creamos un boton con la siguiente etiqueta
button = new Button.with_label ("grabar")
button2 = new Button.with_label ("parar")
button3 = new Button.with_label ("play")
button4 = new Button.with_label ("pausar")
var box= new Box (Gtk.Orientation.HORIZONTAL, 4)
var box1= new Box (Gtk.Orientation.HORIZONTAL, 4)
var box2= new Box (Gtk.Orientation.VERTICAL, 4)
scale_1= new Scale.with_range (Orientation.HORIZONTAL,0,1000,0.1)
scale_1.value_changed.connect(on_scale)
this.drawing_area = new DrawingArea ();
this.drawing_area.realize.connect(on_realize);
// Une el evento de clic de raton con la funcion pulsado
button.clicked.connect (on_grabar)
button2.clicked.connect (on_parar)
button4.clicked.connect (on_pausa)
button3.clicked.connect (on_play)
// si pulsamos la x de la barra saldrá del loop
destroy.connect(Gtk.main_quit)
// añade el boton a la ventana
this.add(box2)
box.add(button3)
box.add(button2)
box.add(button4)
box.add(button)
box2.pack_start (drawing_area, true, true, 0);
box2.pack_start (box,false, false, 0);
box2.pack_start (scale_1,false, true, 0);
estado="STOP"
bus = this.videopipeline.get_bus()
bus.add_signal_watch()
msg = bus.timed_pop_filtered (10,Gst.MessageType.STATE_CHANGED | Gst.MessageType.ERROR | Gst.MessageType.EOS );
bus.message.connect(on_msg)
def on_msg(m:Gst.Message)
if m.type== Gst.MessageType.STATE_CHANGED
old_state:Gst.State;
new_state:Gst.State;
pending_state:Gst.State;
m.parse_state_changed (out old_state, out new_state, out pending_state);
if (m.src == this.videosink)
// Remember whether we are in the PLAYING state or not:
if (new_state == Gst.State.PLAYING)
q : Gst.Query = new Gst.Query.seeking (Gst.Format.TIME);
start:int64;
end:int64;
if ( this.videosink.query (q) )
q.parse_seeking (null, out this.seek_enabled, out start, out end);
if seek_enabled
pass
if (new_state == Gst.State.READY)
//print "STOP"
pass
if (new_state == Gst.State.PAUSED)
print "PAUSE"
q : Gst.Query = new Gst.Query.seeking (Gst.Format.TIME);
start:int64;
end:int64;
if ( this.videosink.query (q) )
q.parse_seeking (null, out this.seek_enabled, out start, out end);
if seek_enabled
//print "enable cuando pause"+start.to_string()+"-"+end.to_string()
pass
def on_scale( )//cuando la escala se mueve por el usuario
if seek_enabled
if this.videopipeline.seek_simple(Gst.Format.TIME, SeekFlags.FLUSH| Gst.SeekFlags.ACCURATE, (int64)(scale_1.get_value()*Gst.SECOND))
print "moviendo video"
else
print "video no se puede mover"
def mover():bool // cambiando el valor mientras play el video.
//desconectamos la deteccion de valores cambiados de la escala para evitar que el programa crea que el usuario esta cambiando el valor.
scale_1.value_changed.disconnect(on_scale)
var format = Gst.Format.TIME
position=0
if this.videopipeline.query_position(ref format, out position)
scale_1.set_value(position/Gst.MSECOND/1000)
duracion=0;
if this.videopipeline.query_duration(ref format, out duracion)
duracion=duracion/Gst.MSECOND/1000
if estado!="STOP" do scale_1.set_range(0,duracion)
scale_1.value_changed.connect(on_scale)
return true
def on_realize()
this.xid = (ulong)Gdk.X11Window.get_xid(this.drawing_area.get_window());
def on_grabar (btn : Button)
if estado=="STOP" or estado=="PAUSA"
recordsink.set("location","grabacion"+(numgrab).to_string()+".wav")
this.videopipeline.add(this.recordbin)
this.videopipeline.remove(audiobin)
print "grabando"
button.set_label("parar")
var xoverlay = this.videosink as XOverlay;
xoverlay.set_xwindow_id (this.xid);
this.videopipeline.set_state (State.PLAYING);
estado="REC"
var format = Gst.Format.TIME
position: int64
if this.videopipeline.query_position(ref format, out position)
comienzo_grabacion=(position/Gst.MSECOND/1000)
else
comienzo_grabacion=-1
else if estado=="REC"
//parando la grabación
estado="PLAY"
button.set_label("grabar")
videopipeline.set_state(State.PAUSED)
this.videopipeline.remove(recordbin)
this.recordbin.set_state(State.NULL)
this.recordbin.send_event(new Event.eos())
this.videopipeline.add(audiobin)
this.on_insertar()
def on_insertar()
if comienzo_grabacion!=-1
try
Process.spawn_command_line_sync ("sox grabacion"+numgrab.to_string()+".wav grabacion_x.wav pad "+((int64)comienzo_grabacion).to_string()+"@0")
Process.spawn_command_line_sync ("mv grabacion_x.wav grabacion"+numgrab.to_string()+".wav")
pass
except
pass
print "grabacion a añadir :"+numgrab.to_string()
vaudioarchivos.add((new AudioFilesSrc()))
vaudioarchivos.last().open(this.audiobin,this.vaudioadder,"grabacion"+numgrab.to_string()+".wav")
vaudioarchivos.last().conecta(this.vaudioadder)
numgrab+=1
def on_pausa (btn:Button)
if estado=="PLAY"
print "pausando"
this.videopipeline.set_state (State.PAUSED);
estado="PAUSA"
def on_parar (btn : Button)
if estado=="PLAY" or estado=="PAUSA"
var xoverlay = this.videosink as XOverlay;
xoverlay.set_xwindow_id (this.xid);
this.videopipeline.set_state (State.READY);
this.scale_1.set_value(0)
estado="STOP"
def on_play (btn : Button)
if estado=="STOP" or estado=="PAUSA"
print "tocando"
var xoverlay = this.videosink as XOverlay;
xoverlay.set_xwindow_id (this.xid);
this.videopipeline.set_state (State.PLAYING);
estado="PLAY"
def OnDynamicPad (element:Element, zz:Pad)
var opad = this.videosink.get_static_pad("sink");
zz.link(opad);
我对您的代码进行了一些小的修改,并使用以下build命令使用gstreamer-1.0对其进行了编译:
valac --pkg gdk-x11-3.0 --pkg gstreamer-1.0 --pkg gstreamer-video-1.0 --pkg gee-0.8 --pkg gtk+-3.0 bikoizketa.gs
如果我点击“grabar”按钮,视频就会播放,我可以录制自己的讲话。当我点击同一个按钮(现在是“parar”)时,它会将该录音保存到一个名为grabacion1.wav的文件中,该文件播放效果良好,但仅是来自麦克风的录音
我所做的修改主要是将其更新为gstreamer1.0api和新的gdkx11api。我还为刻度(第254行)添加了一个负值检查,因为处于记录状态的流返回-1(可能是因为流持续时间可能是无限的)
我还注意到,在try-catch子句中有pass
语句阻止它抛出错误。删除这些内容后,如果我录制了一些内容,点击“pausar”按钮,然后再次录制,则会出现以下错误:
sox WARN wav:.wav输入文件上的过早EOF
[indent=4]
uses
Gtk
Gst
Gee
class AudioFilesSrc : Gst.Bin
wavparse: Element
src: Element
audioconvert: Element
srcpad: Pad
sinkpad: Pad
def OnDynamicPad (element:Element, zz:Pad)
var opad = audioconvert.get_static_pad("sink");
zz.link(opad);
def open(p:Gst.Bin,mixer:Element,s1:string)
src = Gst.ElementFactory.make("filesrc", "src1");
wavparse = ElementFactory.make("wavparse","wavparse");
audioconvert = ElementFactory.make("audioconvert","audioconvert");
wavparse.pad_added.connect(OnDynamicPad);
this.add_many(src,wavparse,audioconvert);
src.link_many(wavparse,audioconvert);
src.set("location",s1);
// añade este bin a pipeline general
p.add(this)
//busca la salida de audioconvert1 y conviertela en la salida del bin
srcpad = new Gst.GhostPad("src", audioconvert.get_static_pad("src"));
this.add_pad(srcpad);
print ".-.----------------------------- abierto "+s1
def conecta(mixer:Element)
//crea una entrada al mixer
sinkpad = mixer.get_request_pad("sink_%u")
//this.get_pad("src").link(this.sinkpad)
srcpad.link(this.sinkpad)
//this.srcpad.set_blocked(false)
print ".-.----------------------------- conectado"
def close(p:Gst.Bin,mixer:Element)
p.set_state(State.PAUSED)
//this.srcpad.set_blocked(true)
this.srcpad.add_probe(PadProbeType.BLOCK, cb)
this.set_state(State.NULL)
this.srcpad.unlink(this.sinkpad)
mixer.release_request_pad (this.sinkpad)
def cb (p:Pad, info:PadProbeInfo) : PadProbeReturn
return PadProbeReturn.OK
init
Gtk.init (ref args)
Gst.init (ref args);
var prueba = new ventana ()
prueba.show_all ()
Gtk.main ();
class ventana : Window
drawing_area:private DrawingArea;
videopipeline: private Pipeline ;
recordbin: Gst.Bin
videobin: Gst.Bin
audiobin: Gst.Bin
volume: private Element ;
videosrc :private Element;
videosink: private Element;
videodec: private Element;
vaudiosink: private Element;
vaudioparser: Element;
vaudiodec: Element;
vaudioadder: private Element;
vaudioarchivos:list of AudioFilesSrc
recordsrc :private Element;
recordsink: private Element;
recordconvert: private Element;
recordencoder: private Element;
comienzo_grabacion:float
xid :private ulong ;
reloj:uint
position: float
duracion:float
bus:Gst.Bus
bus2:Gst.Bus
msg:Gst.Message
msg2:Gst.Message
seek_enabled:bool
seek_enabled2:bool
scale_1:Scale
estado:string
lugar:int
numgrab:int=0
archivos_audio:list of string
button:Button
button1:Button
button2:Button
button3:Button
button4:Button
button5:Button
button6:Button
button7:Button
button8:Button
init
reloj = Timeout.add(1000, mover)
title = "Bikoizketa"
default_height = 250
default_width = 450
window_position = WindowPosition.CENTER
comienzo_grabacion=-1
// video pipeline
duracion=Gst.CLOCK_TIME_NONE;
this.videopipeline = new Pipeline ("mypipeline");
videobin= new Gst.Bin("videobin")
this.videosrc = ElementFactory.make ("filesrc", "filesrc2")
this.videosrc.set("location","gontzal3.mp4");
this.videodec = ElementFactory.make ("decodebin", "dec");
this.videosink = ElementFactory.make ("xvimagesink", "videosink");
this.videosink.set("force-aspect-ratio",true)
this.videodec.pad_added.connect(OnDynamicPad);
this.videobin.add_many (videosrc,videodec,videosink)
this.videosrc.link_many (videodec,videosink)
this.videopipeline.add(videobin)
this.audiobin= new Gst.Bin("audiobin")
this.vaudioadder = ElementFactory.make("adder","mixer");
this.volume= ElementFactory.make("volume","volume");
this.volume.set_property("volume",0.5)
// creamos un elemento de autoaudiosink (en lugar de alsaaudiosink)
this.vaudiosink= ElementFactory.make("autoaudiosink","autoaudiosink");
this.audiobin.add_many (vaudioadder,volume,vaudiosink)
vaudioadder.link_many(volume,vaudiosink)
this.videopipeline.add(audiobin)
//añadiendo archivos
vaudioarchivos=new list of AudioFilesSrc
vaudioarchivos.add (new AudioFilesSrc())
vaudioarchivos.last().open(audiobin,vaudioadder,"silencio.wav")
vaudioarchivos.last().conecta(vaudioadder)
numgrab++
this.recordbin= new Gst.Bin("recorder")
// creamos un elemento de autoaudiosrc (en lugar de alsaaudiosrc)
this.recordsrc= ElementFactory.make ("autoaudiosrc","autoaudiosrc")
this.recordconvert=ElementFactory.make ("audioconvert","audioconvert")
this.recordencoder = Gst.ElementFactory.make("wavenc", "encoder")
this.recordsink= ElementFactory.make ("filesink","filesink")
this.recordsink.set ("location","file.wav")
this.recordbin.add_many (this.recordsrc,this.recordconvert, this.recordencoder, this.recordsink);
this.recordsrc.link_many(this.recordconvert,recordencoder,recordsink)
//videopipeline.add(recordbin)
// creamos un boton con la siguiente etiqueta
button = new Button.with_label ("grabar")
button2 = new Button.with_label ("parar")
button3 = new Button.with_label ("play")
button4 = new Button.with_label ("pausar")
var box= new Box (Gtk.Orientation.HORIZONTAL, 4)
var box1= new Box (Gtk.Orientation.HORIZONTAL, 4)
var box2= new Box (Gtk.Orientation.VERTICAL, 4)
scale_1= new Scale.with_range (Orientation.HORIZONTAL,0,1000,0.1)
scale_1.value_changed.connect(on_scale)
this.drawing_area = new DrawingArea ();
this.drawing_area.realize.connect(on_realize);
// Une el evento de clic de raton con la funcion pulsado
button.clicked.connect (on_grabar)
button2.clicked.connect (on_parar)
button4.clicked.connect (on_pausa)
button3.clicked.connect (on_play)
// si pulsamos la x de la barra saldrá del loop
destroy.connect(Gtk.main_quit)
// añade el boton a la ventana
this.add(box2)
box.add(button3)
box.add(button2)
box.add(button4)
box.add(button)
box2.pack_start (drawing_area, true, true, 0);
box2.pack_start (box,false, false, 0);
box2.pack_start (scale_1,false, true, 0);
estado="STOP"
bus = this.videopipeline.get_bus()
bus.add_signal_watch()
msg = bus.timed_pop_filtered (10,Gst.MessageType.STATE_CHANGED | Gst.MessageType.ERROR | Gst.MessageType.EOS );
bus.message.connect(on_msg)
def on_msg(m:Gst.Message)
if m.type== Gst.MessageType.STATE_CHANGED
old_state:Gst.State;
new_state:Gst.State;
pending_state:Gst.State;
m.parse_state_changed (out old_state, out new_state, out pending_state);
if (m.src == this.videosink)
// Remember whether we are in the PLAYING state or not:
if (new_state == Gst.State.PLAYING)
q : Gst.Query = new Gst.Query.seeking (Gst.Format.TIME);
start:int64;
end:int64;
if ( this.videosink.query (q) )
q.parse_seeking (null, out this.seek_enabled, out start, out end);
if seek_enabled
pass
if (new_state == Gst.State.READY)
//print "STOP"
pass
if (new_state == Gst.State.PAUSED)
print "PAUSE"
q : Gst.Query = new Gst.Query.seeking (Gst.Format.TIME);
start:int64;
end:int64;
if ( this.videosink.query (q) )
q.parse_seeking (null, out this.seek_enabled, out start, out end);
if seek_enabled
//print "enable cuando pause"+start.to_string()+"-"+end.to_string()
pass
def on_scale( )//cuando la escala se mueve por el usuario
if seek_enabled
if this.videopipeline.seek_simple(Gst.Format.TIME, SeekFlags.FLUSH| Gst.SeekFlags.ACCURATE, (int64)(scale_1.get_value()*Gst.SECOND))
print "moviendo video"
else
print "video no se puede mover"
def mover():bool // cambiando el valor mientras play el video.
//desconectamos la deteccion de valores cambiados de la escala para evitar que el programa crea que el usuario esta cambiando el valor.
scale_1.value_changed.disconnect(on_scale)
var format = Gst.Format.TIME
position=0
if this.videopipeline.query_position(format, out position)
scale_1.set_value(position/Gst.MSECOND/1000)
duracion=0;
if this.videopipeline.query_duration(format, out duracion)
duracion=duracion >= 0 ? duracion/Gst.MSECOND/1000 : 0
if estado!="STOP" do scale_1.set_range(0,duracion)
scale_1.value_changed.connect(on_scale)
return true
def on_realize()
this.xid = (ulong)((Gdk.X11.Window)this.drawing_area.get_window()).get_xid();
def on_grabar (btn : Button)
if estado=="STOP" or estado=="PAUSA"
recordsink.set("location","grabacion"+(numgrab).to_string()+".wav")
this.videopipeline.add(this.recordbin)
this.videopipeline.remove(audiobin)
print "grabando"
button.set_label("parar")
var xoverlay = this.videosink as Video.Overlay
xoverlay.set_window_handle ((uint*)this.xid);
this.videopipeline.set_state (State.PLAYING);
estado="REC"
var format = Gst.Format.TIME
position: int64
if this.videopipeline.query_position(format, out position)
comienzo_grabacion=(position/Gst.MSECOND/1000)
else
comienzo_grabacion=-1
else if estado=="REC"
//parando la grabación
estado="PLAY"
button.set_label("grabar")
videopipeline.set_state(State.PAUSED)
this.videopipeline.remove(recordbin)
this.recordbin.set_state(State.NULL)
this.recordbin.send_event(new Event.eos())
this.videopipeline.add(audiobin)
this.on_insertar()
def on_insertar()
if comienzo_grabacion!=-1
try
Process.spawn_command_line_sync ("sox grabacion"+numgrab.to_string()+".wav grabacion_x.wav pad "+((int64)comienzo_grabacion).to_string()+"@0")
Process.spawn_command_line_sync ("mv grabacion_x.wav grabacion"+numgrab.to_string()+".wav")
except err:Error
print err.message
print "grabacion a añadir :"+numgrab.to_string()
vaudioarchivos.add((new AudioFilesSrc()))
vaudioarchivos.last().open(this.audiobin,this.vaudioadder,"grabacion"+numgrab.to_string()+".wav")
vaudioarchivos.last().conecta(this.vaudioadder)
numgrab+=1
def on_pausa (btn:Button)
if estado=="PLAY"
print "pausando"
this.videopipeline.set_state (State.PAUSED);
estado="PAUSA"
def on_parar (btn : Button)
if estado=="PLAY" or estado=="PAUSA"
var xoverlay = this.videosink as Video.Overlay
xoverlay.set_window_handle ((uint*)this.xid);
this.videopipeline.set_state (State.READY);
this.scale_1.set_value(0)
estado="STOP"
def on_play (btn : Button)
if estado=="STOP" or estado=="PAUSA"
print "tocando"
var xoverlay = this.videosink as Video.Overlay
xoverlay.set_window_handle ((uint*)this.xid);
this.videopipeline.set_state (State.PLAYING);
estado="PLAY"
def OnDynamicPad (element:Element, zz:Pad)
var opad = this.videosink.get_static_pad("sink");
zz.link(opad);
如果我理解正确,您最终会希望录制该人每次按下“grabar”时的讲话,并将其保存到grabacion1.wav、grabacion2.wav等文件序列中,然后在他们完成将其与视频组合并将输出保存到文件中时?如果是这样,我不太明白最后一点应该发生在哪里。Kaixo!您使用gstreamer 0.1而不是gstreamer 1.0有什么原因吗?如果我使用gstreamer 1.0编译代码(以及一些小的修改),它将运行并生成音频文件,而不会崩溃。当终端崩溃时,您在终端中得到任何输出吗?Kaixo!!!:->唯一的原因是Tarnyko提供的用于交叉编译的Windows包。但如果我使用1.0,这不是问题。是的,它产生音频文件,但你能一起听吗?。还没有输出。我尝试使用1.0,在运行def conecta:GStreamer CRITICAL**:gst\u pad\u link\u full:assertion'gst\u IS\u pad(sinkpad)'失败,播放后:gtk\u range\u set\u range:assertion'min是的,您已正确解释。录音机保存了版权。但是我不能控制加法器元素。