
Recherche avancée
Autres articles (60)
-
Encodage et transformation en formats lisibles sur Internet
10 avril 2011MediaSPIP transforme et ré-encode les documents mis en ligne afin de les rendre lisibles sur Internet et automatiquement utilisables sans intervention du créateur de contenu.
Les vidéos sont automatiquement encodées dans les formats supportés par HTML5 : MP4, Ogv et WebM. La version "MP4" est également utilisée pour le lecteur flash de secours nécessaire aux anciens navigateurs.
Les documents audios sont également ré-encodés dans les deux formats utilisables par HTML5 :MP3 et Ogg. La version "MP3" (...) -
Submit bugs and patches
13 avril 2011Unfortunately a software is never perfect.
If you think you have found a bug, report it using our ticket system. Please to help us to fix it by providing the following information : the browser you are using, including the exact version as precise an explanation as possible of the problem if possible, the steps taken resulting in the problem a link to the site / page in question
If you think you have solved the bug, fill in a ticket and attach to it a corrective patch.
You may also (...) -
Les autorisations surchargées par les plugins
27 avril 2010, parMediaspip core
autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs
Sur d’autres sites (8250)
-
Revision a4c0b3531e : Make alt_extra_bits a local variable alt_extra_bits is now only used in a local
17 avril 2013, par Adrian GrangeChanged Paths : Modify /vp9/encoder/vp9_firstpass.c Make alt_extra_bits a local variable alt_extra_bits is now only used in a local context so remove it from the twopass_rc structure. Change-Id : I5bbf0a3dba9712a3da45760f7bb865243705b53e
-
Cloaked Archive Wiki
16 mai 2011, par Multimedia Mike — GeneralGoogle’s Chrome browser has made me phenomenally lazy. I don’t even attempt to type proper, complete URLs into the address bar anymore. I just type something vaguely related to the address and let the search engine take over. I saw something weird when I used this method to visit Archive Team’s site :
There’s greater detail when you elect to view more results from the site :
As the administrator of a MediaWiki installation like the one that archiveteam.org runs on, I was a little worried that they might have a spam problem. However, clicking through to any of those out-of-place pages does not indicate anything related to pharmaceuticals. Viewing source also reveals nothing amiss.
I quickly deduced that this is a textbook example of website cloaking. This is when a website reports different content to a search engine than it reports to normal web browsers (humans, presumably). General pseudocode :
C :-
if (web_request.user_agent_string == CRAWLER_USER_AGENT)
-
return cloaked_data ;
-
else
-
return real_data ;
You can verify this for yourself using the
wget
command line utility :<br />
$ wget --quiet --user-agent="<strong>Mozilla/5.0</strong>" \<br />
http://www.archiveteam.org/index.php?title=Geocities -O - | grep \<title\><br />
<title>GeoCities - Archiveteam</title>$ wget —quiet —user-agent="Googlebot/2.1"
http://www.archiveteam.org/index.php?title=Geocities -O - | grep \<title\>
<title>Cheap xanax | Online Drug Store, Big Discounts</title>I guess the little web prank worked because the phaux-pharma stuff got indexed. It makes we wonder if there’s a MediaWiki plugin that does this automatically.
For extra fun, here’s a site called the CloakingDetector which purports to be able to detect whether a page employs cloaking. This is just one humble observer’s opinion, but I don’t think the site works too well :
-
-
avcodec/hevc : fix no output of prior pics and pic output flags(cherry picked from...
29 avril 2014, par gcocherel