Include a User-Agent header equivalent to the cURL one; this also is needed when fetching export pages from Wikipedia, as we currently block the default PHP user-agent string...
116 lines
3.1 KiB
PHP
116 lines
3.1 KiB
PHP
<?php
|
|
|
|
/**
|
|
* Various HTTP related functions
|
|
*/
|
|
class Http {
|
|
static function get( $url, $timeout = 'default' ) {
|
|
return Http::request( "GET", $url, $timeout );
|
|
}
|
|
|
|
static function post( $url, $timeout = 'default' ) {
|
|
return Http::request( "POST", $url, $timeout );
|
|
}
|
|
|
|
/**
|
|
* Get the contents of a file by HTTP
|
|
*
|
|
* if $timeout is 'default', $wgHTTPTimeout is used
|
|
*/
|
|
static function request( $method, $url, $timeout = 'default' ) {
|
|
global $wgHTTPTimeout, $wgHTTPProxy, $wgVersion, $wgTitle;
|
|
|
|
# Use curl if available
|
|
if ( function_exists( 'curl_init' ) ) {
|
|
$c = curl_init( $url );
|
|
if ( wfIsLocalURL( $url ) ) {
|
|
curl_setopt( $c, CURLOPT_PROXY, 'localhost:80' );
|
|
} else if ($wgHTTPProxy) {
|
|
curl_setopt($c, CURLOPT_PROXY, $wgHTTPProxy);
|
|
}
|
|
|
|
if ( $timeout == 'default' ) {
|
|
$timeout = $wgHTTPTimeout;
|
|
}
|
|
curl_setopt( $c, CURLOPT_TIMEOUT, $timeout );
|
|
curl_setopt( $c, CURLOPT_USERAGENT, "MediaWiki/$wgVersion" );
|
|
if ( $method == 'POST' )
|
|
curl_setopt( $c, CURLOPT_POST, true );
|
|
else
|
|
curl_setopt( $c, CURLOPT_CUSTOMREQUEST, $method );
|
|
|
|
# Set the referer to $wgTitle, even in command-line mode
|
|
# This is useful for interwiki transclusion, where the foreign
|
|
# server wants to know what the referring page is.
|
|
# $_SERVER['REQUEST_URI'] gives a less reliable indication of the
|
|
# referring page.
|
|
if ( is_object( $wgTitle ) ) {
|
|
curl_setopt( $c, CURLOPT_REFERER, $wgTitle->getFullURL() );
|
|
}
|
|
|
|
ob_start();
|
|
curl_exec( $c );
|
|
$text = ob_get_contents();
|
|
ob_end_clean();
|
|
|
|
# Don't return the text of error messages, return false on error
|
|
if ( curl_getinfo( $c, CURLINFO_HTTP_CODE ) != 200 ) {
|
|
$text = false;
|
|
}
|
|
curl_close( $c );
|
|
} else {
|
|
# Otherwise use file_get_contents...
|
|
# This may take 3 minutes to time out, and doesn't have local fetch capabilities
|
|
|
|
global $wgVersion;
|
|
$headers = array( "User-Agent: MediaWiki/$wgVersion" );
|
|
if( strcasecmp( $method, 'post' ) == 0 ) {
|
|
// Required for HTTP 1.0 POSTs
|
|
$headers[] = "Content-Length: 0";
|
|
}
|
|
$opts = array(
|
|
'http' => array(
|
|
'method' => $method,
|
|
'header' => implode( "\r\n", $headers ) ) );
|
|
$ctx = stream_context_create($opts);
|
|
|
|
$url_fopen = ini_set( 'allow_url_fopen', 1 );
|
|
$text = file_get_contents( $url, false, $ctx );
|
|
ini_set( 'allow_url_fopen', $url_fopen );
|
|
}
|
|
return $text;
|
|
}
|
|
|
|
/**
|
|
* Check if the URL can be served by localhost
|
|
*/
|
|
static function isLocalURL( $url ) {
|
|
global $wgCommandLineMode, $wgConf;
|
|
if ( $wgCommandLineMode ) {
|
|
return false;
|
|
}
|
|
|
|
// Extract host part
|
|
$matches = array();
|
|
if ( preg_match( '!^http://([\w.-]+)[/:].*$!', $url, $matches ) ) {
|
|
$host = $matches[1];
|
|
// Split up dotwise
|
|
$domainParts = explode( '.', $host );
|
|
// Check if this domain or any superdomain is listed in $wgConf as a local virtual host
|
|
$domainParts = array_reverse( $domainParts );
|
|
for ( $i = 0; $i < count( $domainParts ); $i++ ) {
|
|
$domainPart = $domainParts[$i];
|
|
if ( $i == 0 ) {
|
|
$domain = $domainPart;
|
|
} else {
|
|
$domain = $domainPart . '.' . $domain;
|
|
}
|
|
if ( $wgConf->isLocalVHost( $domain ) ) {
|
|
return true;
|
|
}
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
}
|
|
|