Recherche avancée

Médias (91)

Autres articles (84)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Support de tous types de médias

    10 avril 2011

    Contrairement à beaucoup de logiciels et autres plate-formes modernes de partage de documents, MediaSPIP a l’ambition de gérer un maximum de formats de documents différents qu’ils soient de type : images (png, gif, jpg, bmp et autres...) ; audio (MP3, Ogg, Wav et autres...) ; vidéo (Avi, MP4, Ogv, mpg, mov, wmv et autres...) ; contenu textuel, code ou autres (open office, microsoft office (tableur, présentation), web (html, css), LaTeX, Google Earth) (...)

  • MediaSPIP v0.2

    21 juin 2013, par

    MediaSPIP 0.2 is the first MediaSPIP stable release.
    Its official release date is June 21, 2013 and is announced here.
    The zip file provided here only contains the sources of MediaSPIP in its standalone version.
    To get a working installation, you must manually install all-software dependencies on the server.
    If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...)

Sur d’autres sites (8072)

  • Started Programming Young

    6 septembre 2011, par Multimedia Mike — Programming

    I have some of the strangest memories of my struggles to jump into computer programming.

    Back To BASIC
    I remember doing some Logo programming on Apple II computers at school in 5th grade (1987 timeframe). But that was mostly driving turtle graphics. Then I remember doing some TRS-80 BASIC in 7th grade, circa 1989. Emboldened by what very little I had learned in perhaps the week or 2 we took in a science class to do this, I tried a little GW-BASIC on my family’s “IBM-PC compatible” computer (they were still called that back then). I still remember what my first program consisted of. Even back then I was interested in manipulating graphics and color on a computer screen. Thus :

    10 color 1
    20 print "This is color 1"
    30 color 2
    40 print "This is color 2"
    ...
    

    And so on through 15 colors. Hey, it did the job– it demonstrated the 15 different colors you could set in text mode.

    What’s FOR For ?
    That 7th grade computer unit in science class wasn’t very thick on computer science details. I recall working with a lab partner to transcribe code listings into a computer (and also saving my work to a storage cassette). We also developed form processing programs that would print instructions to input text followed by an “INPUT I$” statement to obtain the user’s output.

    I remember there was some situation where we needed a brief delay between input and printing. The teacher told us to use a construct of the form :

    10 FOR I = 1 TO 20000
    20 NEXT I
    

    We had to calibrate the number based on our empirical assessment of how long it lasted but I recall that the number couldn’t be much higher than about 32000, for reasons that would become clearer much later.

    Imagine my confusion when I would read and try to comprehend BASIC program code I would find in magazines. I would of course see that FOR..NEXT construct all over the place but obviously not in the context of introducing deliberate execution delays. Indeed, my understanding of one of the fundamental building blocks of computer programming — iteration — was completely skewed because of this early lesson.

    Refactoring
    Somewhere along the line, I figured out that the FOR..NEXT could be used to do the same thing a bunch of times, possibly with different values. A few years after I had written that color program, I found it again and realized that I could write it as :

    10 for I = 1 to 15
    20 color I
    30 print I
    40 next I
    

    It still took me a few more years to sort out the meaning of WHILE..WEND, though.

  • avcodec/mpegvideo_dec : Don't use MotionEstContext as scratch space

    30 octobre 2022, par Andreas Rheinhardt
    avcodec/mpegvideo_dec : Don't use MotionEstContext as scratch space
    

    Decoders that might use quarter pixel motion estimation
    (namely MPEG-4 as well as the VC-1 family) currently
    use MpegEncContext.me.qpel_(put|avg) as scratch space
    for pointers to arrays of function pointers.
    (MotionEstContext contains such pointers as it supports
    quarter pixel motion estimation.) The MotionEstContext
    is unused apart from this for the decoding part of
    mpegvideo.

    Using the context at all is for decoding is actually
    unnecessary and easily avoided : All codecs with
    quarter pixels set me.qpel_avg to qdsp.avg_qpel_pixels_tab,
    so one can just unconditionally use this in ff_mpv_reconstruct_mb().
    MPEG-4 sets qpel_put to qdsp.put_qpel_pixels_tab
    or to qdsp.put_no_rnd_qpel_pixels_tab based upon
    whether the current frame is a b-frame with no_rounding
    or not, while the VC-1-based decoders set it to
    qdsp.put_qpel_pixels_tab unconditionally. Given
    that no_rounding is always zero for VC-1, using
    the same check for VC-1 as well as for MPEG-4 would work.
    Since ff_mpv_reconstruct_mb() already has exactly
    the right check (for hpeldsp), it can simply be reused.

    (This change will result in ff_mpv_motion() receiving
    a pointer to an array of NULL function pointers instead
    of a NULL pointer for codecs without qpeldsp (like MPEG-1/2).
    It doesn't matter.)

    Signed-off-by : Andreas Rheinhardt <andreas.rheinhardt@outlook.com>

    • [DH] libavcodec/h263dec.c
    • [DH] libavcodec/mpv_reconstruct_mb_template.c
    • [DH] libavcodec/mss2.c
    • [DH] libavcodec/vc1dec.c
  • Stuck in installing a voicecloner via Python (module not found)

    25 novembre 2023, par Wimmah

    I use Python 3.11.5

    &#xA;

    As a great Python n00b I enter this forum because I'm stuck with installing a Voice Cloner (for personal use to do a funny trick for X-mas with my family) Its this tool that i'm trying to install : https://github.com/CorentinJ/Real-Time-Voice-Cloning

    &#xA;

    With a little help of chatGTP I came quite far but for some reason the downloaded datasets cant be found. Instructions of the tool state :

    &#xA;

    Install intructions form Github&#xA;So my tree looks like this :

    &#xA;

    (base) willem@willems-air Voice cloner % tree&#xA;.&#xA;├── demo_cli.py&#xA;├── demo_toolbox.py&#xA;├── encoder_preprocess.py&#xA;├── encoder_train.py&#xA;├── saved_models&#xA;│&#xA0;&#xA0; └── default&#xA;│&#xA0;&#xA0;     ├── encoder.pt&#xA;│&#xA0;&#xA0;     ├── synthesizer.pt&#xA;│&#xA0;&#xA0;     └── vocoder.pt&#xA;├── synthesizer_preprocess_audio.py&#xA;├── synthesizer_preprocess_embeds.py&#xA;├── synthesizer_train.py&#xA;└── vocoder_train.py&#xA;&#xA;3 directories, 11 files&#xA;

    &#xA;

    However, when I give the command to execute the demo, I get the message that a needed module cant be found :

    &#xA;

    (base) willem@willems-air Voice cloner % python demo_cli.py&#xA;Traceback (most recent call last):&#xA;  File "/Users/willem/Desktop/Voice cloner/demo_cli.py", line 10, in <module>&#xA;    from encoder import inference as encoder&#xA;ModuleNotFoundError: No module named &#x27;encoder&#x27;&#xA;</module>

    &#xA;

    I build a tree that (for me) looks inline with the installation instructions...(And of course i downloaded the modules without any errors)&#xA;Here also the first lines of the command demo_cli.py where you also see the path :

    &#xA;

    import argparse&#xA;import os&#xA;from pathlib import Path&#xA;&#xA;import librosa&#xA;import numpy as np&#xA;import soundfile as sf&#xA;import torch&#xA;&#xA;from encoder import inference as encoder&#xA;from encoder.params_model import model_embedding_size as speaker_embedding_size&#xA;from synthesizer.inference import Synthesizer&#xA;from utils.argutils import print_args&#xA;from utils.default_models import ensure_default_models&#xA;from vocoder import inference as vocoder&#xA;&#xA;&#xA;if __name__ == &#x27;__main__&#x27;:&#xA;    parser = argparse.ArgumentParser(&#xA;        formatter_class=argparse.ArgumentDefaultsHelpFormatter&#xA;    )&#xA;    parser.add_argument("-e", "--enc_model_fpath", type=Path,&#xA;                        default="saved_models/default/encoder.pt",&#xA;

    &#xA;

    I think i missed out a quite basic step here, but this far ChatGTP is looping and cant help any more, so I need a human tip i guess ;)

    &#xA;

    Thx in advance !

    &#xA;