
Recherche avancée
Autres articles (95)
-
MediaSPIP 0.1 Beta version
25 avril 2011, parMediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
The zip file provided here only contains the sources of MediaSPIP in its standalone version.
To get a working installation, you must manually install all-software dependencies on the server.
If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...) -
Keeping control of your media in your hands
13 avril 2011, parThe vocabulary used on this site and around MediaSPIP in general, aims to avoid reference to Web 2.0 and the companies that profit from media-sharing.
While using MediaSPIP, you are invited to avoid using words like "Brand", "Cloud" and "Market".
MediaSPIP is designed to facilitate the sharing of creative media online, while allowing authors to retain complete control of their work.
MediaSPIP aims to be accessible to as many people as possible and development is based on expanding the (...) -
Participer à sa traduction
10 avril 2011Vous pouvez nous aider à améliorer les locutions utilisées dans le logiciel ou à traduire celui-ci dans n’importe qu’elle nouvelle langue permettant sa diffusion à de nouvelles communautés linguistiques.
Pour ce faire, on utilise l’interface de traduction de SPIP où l’ensemble des modules de langue de MediaSPIP sont à disposition. ll vous suffit de vous inscrire sur la liste de discussion des traducteurs pour demander plus d’informations.
Actuellement MediaSPIP n’est disponible qu’en français et (...)
Sur d’autres sites (9082)
-
ffmpeg get RGB values ?
13 novembre 2012, par NavI've seen this, this and this but I still can't figure out how to get the RGB values from the tutorial code.
if(avcodec_open(pCodecCtx, pCodec)<0) {_getch();return -1;}
pFrame = avcodec_alloc_frame();
pFrameRGB=avcodec_alloc_frame();
if(pFrameRGB==NULL) {_getch();return -1;}
numBytes=avpicture_get_size(PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height);
buffer=(uint8_t *)av_malloc(numBytes*sizeof(uint8_t));
// Assign appropriate parts of buffer to image planes in pFrameRGB
avpicture_fill((AVPicture *)pFrameRGB, buffer, PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height);
i=0;
int framecounter=0;
while(av_read_frame(pFormatCtx, &packet)>=0)
{
if(packet.stream_index==videoStream)
{
// Decode video frame
//int avcodec_decode_video2(AVCodecContext *avctx, AVFrame *picture, int *got_picture_ptr, const AVPacket *avpkt);
avcodec_decode_video2(pCodecCtx, pFrame, &frameFinished, &packet);
if(frameFinished)
{
// Convert the image from its native format to RGB
//img_convert((AVPicture *)pFrameRGB, PIX_FMT_RGB24, (AVPicture*)pFrame, pCodecCtx->pix_fmt, pCodecCtx->width, pCodecCtx->height);
if(++i<=5) {SaveFrame(pFrameRGB, pCodecCtx->width, pCodecCtx->height, i);}
}
}
av_free_packet(&packet);
}Since
img_convert
wasn't available, I took it from here but it threw an unhandled exception error. Why is the RGB format so tough to understand ? Why assign zero to G and B of RGB, as shown here ?How can I get my RGB values as a simple R, G and B which spans between 0 and 255 ? Also, I don't understand the use of
x
andy
, when video width and height are already given.y
is an increment of height, but what aboutx
?Can anyone please show me a complete working code of obtaining rgb values from a video or at least explain how
data
andlinesize
help in storing RGB info ? The lack of tutorials is appalling and the complexity of the API is surprising, after having worked on Processing videos. -
libavcodec decode AVFrames to FIFO buffer
6 novembre 2012, par user1175197My aim is to decode multiple frames of a video file, accumulate the decoded frames into a FIFO buffer and read them later on. I decode the packet to my AVFRame mFrame :
avcodec_decode_video2(mCodecContext,mFrame,&frameFinished,&mPacket) ;
Normally I can just copy the YUV frames from the mFrame->data[n][0] to my FIFO buffer but I am just trying to reduce the memcpy 's as much as possible. So instead of copying mFrame->data[n][0] I just want to store the mFrame (which is much smaller than the frames it points to) in the buffer and when it comes to reading I can just fetch it and reach the data.
I tried to do this but it did not work. The AVFrames are fetched from the buffer but when you show them on the screen the video is like frozen. You may think that I am using the same mFrame and overwriting it each time I decode a packet but I am not. I am creating a new AVFrame* each time in the decode loop.
Is this problem related to how avcodec works ? Any ideas ?
Thanks
mike -
How do I correctly convert .avi to .flv with ffmpeg ? [closed]
25 septembre 2012, par terbooterUPDATE
Shame on me )
I chacked red5 logs again and found that I placed converted files to wrong place.
Now all works fineI have two red5 apps.
-
Recorder. It can record live stream and save it to flv file to disk
private void startRecord(String uid, String name, IConnection connection) {
// Get a reference to the current broadcast stream.
ClientBroadcastStream stream = (ClientBroadcastStream) this.getBroadcastStream(
connection.getScope(), name);
try {
// Save the stream to disk.
String path = uid + "/" + name;
stream.saveAs(path, true);
System.out.println("file..:" + stream.getSaveFilename());
} catch (Exception e) {
System.out.println("Error while saving stream: " + name + e);
}
}
private void stopRecord(String name) {
IConnection conn = Red5.getConnectionLocal();
System.out.println("Stop recording show for: {}" + conn.getScope().getContextPath());
ClientBroadcastStream stream = (ClientBroadcastStream) this.getBroadcastStream(conn.getScope(), name);
// Stop recording.
if (stream != null) {
stream.stopRecording();
}
} -
Second red5 app (Chat) streams recorded flv file to flash client
public static String serverStreamCreate(String path, String streamName) {
IServerStream serverStream = StreamUtils.createServerStream(Red5.getConnectionLocal().getScope(), streamName);
SimplePlayItem item = SimplePlayItem.build(path);
IPlaylistController controller = new MyPlayListController();
serverStream.setPlaylistController(controller);
serverStream.setRepeat(false);
serverStream.addItem(item);
serverStream.addItem(item);
serverStream.start();
return streamName;
}
If I record stream from flash client to flv file with Recorder and after that stream this flv file back to client with Chat, all works fine.
Now I want to convert avi file to flv and stream it from red5 app to flash client.
I used ffmpegffmpeg -i 24.avi -ar 22050 -an -f flv -b 500k -s 320x240 -y 24_c.flv
But if I stream 24_c.flv from Chat app flash client have no video.
24_c.flv cant be playd by VLC player and have same code information as flv file created by Recorder red5 app.I really dont know where to dig.
-