
Recherche avancée
Médias (91)
-
999,999
26 septembre 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Audio
-
The Slip - Artworks
26 septembre 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Texte
-
Demon seed (wav version)
26 septembre 2011, par
Mis à jour : Avril 2013
Langue : English
Type : Audio
-
The four of us are dying (wav version)
26 septembre 2011, par
Mis à jour : Avril 2013
Langue : English
Type : Audio
-
Corona radiata (wav version)
26 septembre 2011, par
Mis à jour : Avril 2013
Langue : English
Type : Audio
-
Lights in the sky (wav version)
26 septembre 2011, par
Mis à jour : Avril 2013
Langue : English
Type : Audio
Autres articles (79)
-
Le profil des utilisateurs
12 avril 2011, parChaque utilisateur dispose d’une page de profil lui permettant de modifier ses informations personnelle. Dans le menu de haut de page par défaut, un élément de menu est automatiquement créé à l’initialisation de MediaSPIP, visible uniquement si le visiteur est identifié sur le site.
L’utilisateur a accès à la modification de profil depuis sa page auteur, un lien dans la navigation "Modifier votre profil" est (...) -
Configurer la prise en compte des langues
15 novembre 2010, parAccéder à la configuration et ajouter des langues prises en compte
Afin de configurer la prise en compte de nouvelles langues, il est nécessaire de se rendre dans la partie "Administrer" du site.
De là, dans le menu de navigation, vous pouvez accéder à une partie "Gestion des langues" permettant d’activer la prise en compte de nouvelles langues.
Chaque nouvelle langue ajoutée reste désactivable tant qu’aucun objet n’est créé dans cette langue. Dans ce cas, elle devient grisée dans la configuration et (...) -
XMP PHP
13 mai 2011, parDixit Wikipedia, XMP signifie :
Extensible Metadata Platform ou XMP est un format de métadonnées basé sur XML utilisé dans les applications PDF, de photographie et de graphisme. Il a été lancé par Adobe Systems en avril 2001 en étant intégré à la version 5.0 d’Adobe Acrobat.
Étant basé sur XML, il gère un ensemble de tags dynamiques pour l’utilisation dans le cadre du Web sémantique.
XMP permet d’enregistrer sous forme d’un document XML des informations relatives à un fichier : titre, auteur, historique (...)
Sur d’autres sites (14278)
-
ffmpeg. I correct that to insert the logo [closed]
13 mai 2012, par user1390921private void btnStart_Click(object sender, EventArgs e)
{
this.btnStart.Enabled = false;
this.progressBar1.Value = 0;
string srcFile = Path.Combine(this.txtSource.Text, this.lstFiles.SelectedItem.ToString());
MessageBox.Show(srcFile);
string dstFile = Path.Combine(this.txtOutput.Text,
Path.GetFileNameWithoutExtension(this.lstFiles.SelectedItem.ToString())) + "." + this.cboOutputFormat.SelectedItem;
MessageBox.Show(dstFile);
string imafile = Path.Combine("movie= "+this.openFileDialog1.SafeFileName);
MessageBox.Show(imafile);
string videoRateOption = string.Empty;
if (this.cboVideoRate.SelectedIndex != 0)
{
videoRateOption = " -b:v " + this.cboVideoRate.SelectedItem.ToString().Split(' ')[0] + "k ";
MessageBox.Show(videoRateOption);
}
string videoSizeOption = string.Empty;
if (this.lstVideoSize.SelectedIndex != 0)
{
videoSizeOption = " -s " + this.lstVideoSize.SelectedItem.ToString().Split(' ')[0] + " ";
}
this.Text = "Converting...";
ThreadPool.QueueUserWorkItem((object state) =>
{
ConvertFile(srcFile, imafile, dstFile, videoRateOption, videoSizeOption);
});
}
string strFFMPEGOut;
ProcessStartInfo psiProcInfo = new ProcessStartInfo();
TimeSpan estimatedTime = TimeSpan.MaxValue;
StreamReader srFFMPEG;
string ste = ""movie=watermarklogo.png [wm];[in][wm] overlay=main_w-overlay_w-10:main_h-overlay_h-10 [out]"";
string strFFMPEGCmd = " -i "" + srcFile + "" -ar 44100 " + videoRateOption + videoSizeOption +"-vf"+ ste + "-y ""
+ dstFile + """;
psiProcInfo.FileName = Application.StartupPath + ((IntPtr.Size == 8) ? "\x64" : "\x86") + "\ffmpeg.exe";
psiProcInfo.Arguments = strFFMPEGCmd;
psiProcInfo.UseShellExecute = false;
psiProcInfo.WindowStyle = ProcessWindowStyle.Hidden;
psiProcInfo.UseShellExecute = false;
psiProcInfo.RedirectStandardError = true;
psiProcInfo.ErrorDialog = true;
psiProcInfo.RedirectStandardOutput = true;
psiProcInfo.CreateNoWindow = true;
prcFFMPEG.StartInfo = psiProcInfo;
prcFFMPEG.Start();There is no response that I push the start button that I tried to fix the code that I inserted the logo in the movie with Ffmpeg. Is that be okay if I don't choose the path of logo File. and Is there any something wrong that I did.
-
Video Conferencing in HTML5 : WebRTC via Web Sockets
14 juin 2012, par silviaA bit over a week ago I gave a presentation at Web Directions Code 2012 in Melbourne. Maxine and John asked me to speak about something related to HTML5 video, so I went for the new shiny : WebRTC – real-time communication in the browser.
I only had 20 min, so I had to make it tight. I wanted to show off video conferencing without special plugins in Google Chrome in just a few lines of code, as is the promise of WebRTC. To a large extent, I achieved this. But I made some interesting discoveries along the way. Demos are in the slide deck.
UPDATE : Opera 12 has been released with WebRTC support.
Housekeeping : if you want to replicate what I have done, you need to install a Google Chrome Web Browser 19+. Then make sure you go to chrome ://flags and activate the MediaStream and PeerConnection experiment(s). Restart your browser and now you can experiment with this feature. Big warning up-front : it’s not production-ready, since there are still changes happening to the spec and there is no compatible implementation by another browser yet.
Here is a brief summary of the steps involved to set up video conferencing in your browser :
- Set up a video element each for the local and the remote video stream.
- Grab the local camera and stream it to the first video element.
- (*) Establish a connection to another person running the same Web page.
- Send the local camera stream on that peer connection.
- Accept the remote camera stream into the second video element.
Now, the most difficult part of all of this – believe it or not – is the signalling part that is required to build the peer connection (marked with (*)). Initially I wanted to run completely without a server and just enter the remote’s IP address to establish the connection. This is, however, not a functionality that the PeerConnection object provides [might this be something to add to the spec ?].
So, you need a server known to both parties that can provide for the handshake to set up the connection. All the examples that I have seen, such as https://apprtc.appspot.com/, use a channel management server on Google’s appengine. I wanted it all working with HTML5 technology, so I decided to use a Web Socket server instead.
I implemented my Web Socket server using node.js (code of websocket server). The video conferencing demo is in the slide deck in an iframe – you can also use the stand-alone html page. Works like a treat.
While it is still using Google’s STUN server to get through NAT, the messaging for setting up the connection is running completely through the Web Socket server. The messages that get exchanged are plain SDP message packets with a session ID. There are OFFER, ANSWER, and OK packets exchanged for each streaming direction. You can see some of it in the below image :
I’m not running a public WebSocket server, so you won’t be able to see this part of the presentation working. But the local loopback video should work.
At the conference, it all went without a hitch (while the wireless played along). I believe you have to host the WebSocket server on the same machine as the Web page, otherwise it won’t work for security reasons.
A whole new world of opportunities lies out there when we get the ability to set up video conferencing on every Web page – scary and exciting at the same time !
-
Using ffmpeg to play file on windows, should use directshow to render the decoded frame [on hold]
22 juillet 2013, par TerryI want to write a video file player using ffmpeg(support many format).
However ffmpeg(/ffplay.exe) using SDL to render the video frame and play the audio sound. I only want to support windows platform. I think the default windows technology/api is more suitalbe for me to render the decoded frame. Do you think so ?
If so, should I use direct show, or just using direct draw for video frame and direct sound for audio frame.
If using direct show, I think I need to wrapper a direct show source filter(/splitter), then pass down the decoded video/audio frame data to render filter.
Please help me which is the best choice ?