
Recherche avancée
Médias (1)
-
Carte de Schillerkiez
13 mai 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Texte
Autres articles (104)
-
Des sites réalisés avec MediaSPIP
2 mai 2011, parCette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page. -
Amélioration de la version de base
13 septembre 2013Jolie sélection multiple
Le plugin Chosen permet d’améliorer l’ergonomie des champs de sélection multiple. Voir les deux images suivantes pour comparer.
Il suffit pour cela d’activer le plugin Chosen (Configuration générale du site > Gestion des plugins), puis de configurer le plugin (Les squelettes > Chosen) en activant l’utilisation de Chosen dans le site public et en spécifiant les éléments de formulaires à améliorer, par exemple select[multiple] pour les listes à sélection multiple (...) -
Menus personnalisés
14 novembre 2010, parMediaSPIP utilise le plugin Menus pour gérer plusieurs menus configurables pour la navigation.
Cela permet de laisser aux administrateurs de canaux la possibilité de configurer finement ces menus.
Menus créés à l’initialisation du site
Par défaut trois menus sont créés automatiquement à l’initialisation du site : Le menu principal ; Identifiant : barrenav ; Ce menu s’insère en général en haut de la page après le bloc d’entête, son identifiant le rend compatible avec les squelettes basés sur Zpip ; (...)
Sur d’autres sites (14905)
-
No audio in the final video when converting webm blobs to mp4 using ffmpeg
28 septembre 2024, par alpeccaI trying to record user camera and microphone and using MediaRecorder to convert the stream to blobs and sending the blobs every 2 second to the backend using websocket. Everything is working fine, but when I checked the final mp4 video in the backend, it doesn't have any audio to it, I try specifying the audio codec, but still no help.


My frontend code :-


const micStream = await navigator.mediaDevices.getUserMedia({ audio: true });

const recorder = new MediaRecorder(stream, {
 mimeType: 'video/webm;codecs=H264',
 videoBitsPerSecond: 8000000,
 audioBitsPerSecond : 8000000
});

recorder.ondataavailable = (e: BlobEvent) => {
 websocket.send(e.data) 
} 
recorder.start(2000);



And here is the backend code :-


@router.websocket("/streamaudio")
async def websocket_endpoint(websocket: WebSocket):
 await manager.connect(websocket)

 recordingFile = os.path.join(os.getcwd(), f"recording_.mp4")

 command = [
 'ffmpeg', 
 '-y',
 '-i', 
 '-', 
 '-codec:v', 
 'copy', 
 '-c:a', 'aac', 
 '-y',
 '-f', 'mp4',
 recordingFile,
 # "-"
 # f'output{queueNumber}.mp4',
 ] 

 
 try:
 while True:
 try:
 
 data = await websocket.receive_bytes()
 
 process.stdin.send(data)
 
 except RuntimeError:
 break 
 except WebSocketDisconnect:
 print(f"Client disconnected: {websocket.client.host}")
 finally:
 manager.disconnect(websocket)
 await process.stdin.aclose()
 await process.wait() 



-
FFmpeg Not Creating or Appending Final HLS Segment When RTMP Stream Ends [closed]
14 août 2024, par ArjitI am transcoding an RTMP stream to HLS using FFmpeg, and I have a problem : when the RTMP stream finishes, the last segment, which does not reach the target duration, is not created as a
.ts
file. Consequently, this last segment is not appended to the playlist file.

command :


fmpeg -i rtmp://localhost:1935/live/xyz -c:v libx264 -preset veryfast -b:v 400k -c:a aac -b:a 64k -s 426x240 -hls_playlist_type event -hls_time 10 -hls_list_size 0 -hls_segment_filename jack/240p_%03d.ts jack/240p.m3u8



I expected that when the RTMP stream ends or when i terminate that FFmpeg process, FFmpeg would create a .ts file for the last segment, even if it is shorter than the target duration (10 seconds), and that this segment would be appended to the playlist (output.m3u8).


-
Use specific channels from specific streams for the final stream using FFMPEG [closed]
9 juin 2024, par Kazu-kunI want to create a 5.1ch audio stream from a 2ch audio stream. However I want to put a negative delay to LFE, SL and SR channels since they will be played from my bluetooth speaker, so I want to make up for the delay by giving it a negative delay.




ffmpeg -itsoffset 0.1 -i input.mkv -i input.mkv


-filter_complex “


[1:a]pan=5.1(side)|FL=FL|FR=FR|LFEp>

[0:a]pan=5.1(side)|FL=FL|FR=FR|LFEp>

[orig5_1][delayed5_1]amerge=inputs=2,pan=5.1(side)|FL=c6|FR=c7|LFE=c2|SL=c3|SR=c4[a]”


-map "[a]" -map 0:v -map 0:s -c:v copy -c:a flac -c:s copy output.mkv




I tried taking the same input twice with the first one being delayed by 0.1 secs. With the delayed input a 5.1 audio is created and with the original another 5.1 audio is created as well. Then the two audio streams are merged with amerge and combined into a single 5.1 audio by taking delayed channels for FL and FR(These will be played from my MacBook) and by taking original channels for LFE, SL and SR(These will be played from my bluetooth JBL). Finally the video and subs streams are taken from the delayed.


The above command gave a "pipe pipe pipe dqoute" error and when I entered the command again it said :




zsh : unknown file attribute : i


zsh : no such file or directory : FL+FR


zsh : unknown file attribute : i


zsh : no such file or directory : FL+FR


zsh : unknown file attribute : i


zsh : command not found : -map




Please give me a command for my requirement !