PHP Doku:: Führt die Unter-Verbindungen des cURL-Handles aus - function.curl-multi-exec.html

Verlauf / Chronik / History: (1) anzeigen

Sie sind hier:
Doku-StartseitePHP-HandbuchFunktionsreferenzSonstige DiensteCURLcURL Funktionencurl_multi_exec

Ein Service von Reinhard Neidl - Webprogrammierung.

cURL Funktionen




(PHP 5)

curl_multi_execFührt die Unter-Verbindungen des cURL-Handles aus


int curl_multi_exec ( resource $mh , int &$still_running )

Verarbeitet alle Handles; diese Methode kann unabhängig davon aufgerufen werden, ob eines der Handles Daten lesen oder schreiben muß.



Ein von curl_multi_init() zurückgegebenes cURL-Multihandle.


Eine Referenz auf ein Flag um festzustellen, ob auf einem oder mehreren Handles noch gearbeitet wird.


Ein cURL-Code der in den cURL Vordefinierten Konstanten definiert ist.


Es werden nur Fehler zurückgegeben, die den Handle-Stack betreffen. Auch wenn CURLM_OK zurückgegeben wird können auf einzelnen Handles Fehler aufgetreten sein.


Beispiel #1 curl_multi_exec()-Beispiel

In diesem Beispiel werden zwei cURL-Handles erstellt, einem Mehrfach-Handle hinzugefügt und anschließend parallel ausgeführt.

// zwei cURL Resourcen erstellen
$ch1 curl_init();
$ch2 curl_init();

// URL und weitere Optionen setzen

// Mehrfach-Handle erstellen
$mh curl_multi_init();

// die zuvor erstellten Handles hinzufügen

$active null;
// Handles ausführen
do {
$mrc curl_multi_exec($mh$active);
} while (

while (
$active && $mrc == CURLM_OK) {
    if (
curl_multi_select($mh) != -1) {
        do {
$mrc curl_multi_exec($mh$active);
        } while (

//close the handles


Siehe auch

11 BenutzerBeiträge:
- Beiträge aktualisieren...
gnuffo1 at gmail dot com
12.12.2010 5:02
Make sure you always remove the individual curl handles from the multi handle. For a CLI script I wrote that fetched the same pages over and over again every few seconds, I had originally omitted doing this, not even realising I had done so. It worked perfectly on my Windows 7 box. However, on my CentOS server, after about 45 seconds, the script kept dying with absolutely no warnings or errors. After hours of debugging I finally realised I wasn't removing the handles from the multi handle and lo and behold, it fixed it.
zhongeren at gmail dot com
20.09.2010 4:55
download all the images given in the URL array in parallel !!

 *@param $picsArr Array [0]=> [url],
 *@$picsArr Array will filled with the image data , you can use the data as you want or just save it in the next step.

function getAllPics(&$picsArr){

count($picsArr)<=0) return false;

$hArr = array();//handle array

foreach($picsArr as $k=>$pic){

$h = curl_init();
curl_setopt($h,CURLOPT_RETURNTRANSFER,1);//return the image value


$mh = curl_multi_init();
$hArr as $k => $h)      curl_multi_add_handle($mh,$h);

$running = null;
$running > 0);

// get the result and save it in the result ARRAY
foreach($hArr as $k => $h){
$picsArr[$k]['data'] = curl_multi_getcontent($h);

//close all the connections
foreach($hArr as $k => $h){
$info = curl_getinfo($h);
$tail = $matches[1];

dirkwybe at gmail dot com
12.12.2009 22:50
If you have problems with client timeouts (browser, wget) - like I did - add some keep alive code.

= array(
"file1" => "[url here]",
"file2" => "[url here]",
"file3" => "[url here]",
"file4" => "[url here]"

$mh = curl_multi_init();
$threads = null;
foreach (
$locations as $name => $url)
$f[$name]=fopen ($name.".xml", "w");
curl_setopt ($c[$name], CURLOPT_FILE, $f[$name]);
curl_setopt ($c[$name], CURLOPT_TIMEOUT,600);
curl_multi_add_handle ($mh,$c[$name]);

$t1 = time();

    if (
time() > $t1 + 2)
"keep-alive" ."<br/>";
$t1 = time();   
while (
$threads > 0);

foreach (
$locations as $name => $url)
fclose ($f[$name]);

jakob dot voss at gbv dot de
14.02.2009 14:10
With the current examples using curl-multi-exec is still not much better then just using file_get_contents. Your script has to needlessly wait at least the maximum of all connection times (instead of the sum). A better solution would be to run the connections as another thread in parallel to other parts of the script:

= new CurlManager();
$handle1 = $cm->register($url1);
$handle2 = $cm->register($url2);

$cm->start();  # start connections

$result1 = $cm->result($handle1); # surely NULL because connection has not been finished

sleep($seconds); # do something else (including use of other CurlManagers!)

$result1 = $cm->result($handle1); # maybe NULL if connection takes longer then $seconds

$cm->finish(); # wait if some connections are still open

$result1 = $cm->result($handle1);
$result2 = $cm->result($handle2);

But I am not sure whether such a CurlManger is possible, especially if you may want to use multiple of them (for instance because some URLs are not known at the start of the script). It's a pain that PHP does not support real threads.
jorge dot hebrard at gmail dot com
25.01.2009 6:40
Based on the below comments, I made a simple class.
See note below code.

* Multiple Curl Handlers
* @author Jorge Hebrard ( )
class curlNode{
    static private
    public function
$new =& self::$listenerList[];
$new['url'] = $url;
$this->callback =& $new;
    * Callbacks needs 3 parameters: $url, $html (data of the url), and $lag (execution time)
public function addListener($callback){
$this->callback['callback'] = $callback;
    * curl_setopt() wrapper. Enjoy!
public function setOpt($key,$value){
$this->callback['opt'][$key] = $value;
    * Request all the created curlNode objects, and invoke associated callbacks.
static public function request(){
//create the multiple cURL handle
$mh = curl_multi_init();
# Setup all curl handles
        # Loop through each created curlNode object.
foreach(self::$listenerList as &$listener){
$url = $listener['url'];
$current =& $ch[];
# Init curl and set default options.
            # This can be improved by creating
$current = curl_init();

curl_setopt($current, CURLOPT_URL, $url);
# Since we don't want to display multiple pages in a single php file, do we?
curl_setopt($current, CURLOPT_HEADER, 0);
curl_setopt($current, CURLOPT_RETURNTRANSFER, 1);
# Set defined options, set through curlNode->setOpt();
if (isset($listener['opt'])){
$listener['opt'] as $key => $value){
curl_setopt($current, $key, $value);
$listener['handle'] = $current;
$listener['start'] = microtime(1);
        } unset(

# Main loop execution
do {
# Exec until there's no more data in this iteration.
            # This function has a bug, it
while(($execrun = curl_multi_exec($mh, $running)) == CURLM_CALL_MULTI_PERFORM);
$execrun != CURLM_OK) break; # This should never happen. Optional line.
            # Get information about the handle that just finished the work.
while($done = curl_multi_info_read($mh)) {
# Call the associated listener
foreach(self::$listenerList as $listener){
# Strict compare handles.
if ($listener['handle'] === $done['handle']) {
# Get content
$html = curl_multi_getcontent($done['handle']);
# Call the callback.
# Remove unnecesary handle (optional, script works without it).
curl_multi_remove_handle($mh, $done['handle']);
# Required, or else we would end up with a endless loop.
            # Without it, even when the connections are over, this script keeps running.
if (!$running) break;
# I don't know what these lines do, but they are required for the script to work.
while (($res = curl_multi_select($mh)) === 0);
            if (
$res === false) break; # Select error, should never happen.
} while (true);

# Finish out our script ;)
$curlGoogle = new curlNode('');
$curlGoogle->setOpt(CURLOPT_HEADER, 0);

$curlMySpace = new curlNode('');


NOTE: I think curl has a bug managing local files, when multiple handles are set.
I tested with a local file, giving me 0.1~ sec. latency.
Then I added another handle, google. With this new addition, my original file had a latency of 1.0~ sec. latency!!
But with external sites (try Yahoo and Google), the latency is not a problem. Google alone gives me 1.7~ sec. latency, and with Yahoo, Google gives me 1.8~ sec. latency.

If anyone can explain, I will be very granted.
7.12.2008 12:13
For anyone trying to dynamically add links to mcurl and removing them as they complete (for example to keep a constant of 5 links downloading), I found that the following code works:

= curl_multi_init();
for(;;) {
$mcRes = curl_multi_exec($mcurl, $mcActive)) == CURLM_CALL_MULTI_PERFORM);
$mcRes != CURLM_OK) break;
$done = curl_multi_info_read($mcurl)) {
// code that parses, adds or removes links to mcurl
// here you should check if all the links are finished and break, or add new links to the loop

You should note the return value of curl_multi_select() is ignored. I found that if you do the curl_multi_exec() loop only if the return value of curl_multi_select() is != -1 then new links added to mcurl are ignored. This may be a bug. Hope this saves someone some time.
Jimmy Ruska
13.09.2008 20:56
> replying to

Just sharing my attempt at it.

> while (($res = curl_multi_select($mh)) === 0) {};
This worked on my windows computer (php 5.2.5) but when I ran the curl program in my new centOS server (php 5.1.6) the function never updates unless curl_multi_exec() is added to the loop, taking away the point of using it to save cycles. curl_multi_select() also allows you to set a timeout but it doesn't seem to help, then again, don't see why people wouldn't use it anyway.

Even when curl_multi_select($mh) does work there's no way to know which of the sockets updated in status or if they've received partial data, completely finished, or just timed out. It's not reliable if you want to remove / add data as soon as things finish. Try calling just or some other website with very little data. curl_multi_select($mh) can send non 0 value a couple of thousand times before finishing. Lazily adding a usleep(25000) or some minimal amount can also help not waste cycles.
9.02.2008 5:04

"When you've added the handles you have for the moment (you can still add new ones at any time), you start the transfers by call curl_multi_perform(3).

curl_multi_perform(3) is asynchronous. It will only execute as little as possible and then return back control to your program. It is designed to never block. If it returns CURLM_CALL_MULTI_PERFORM you better call it again soon, as that is a signal that it still has local data to send or remote data to receive."

So it seems the loop in sample script should look this way:

//execute the handles
do {
    while (
CURLM_CALL_MULTI_PERFORM === curl_multi_exec($mh, $running));
    if (!
$running) break;
    while ((
$res = curl_multi_select($mh)) === 0) {};
    if (
$res === false) {
"<h1>select error</h1>";
} while (

This worked fine (PHP 5.2.5 @ FBSD 6.2) without running non-blocked loop and wasting CPU time.

However this seems to be the only use of curl_multi_select, coz there's no simple way to bind it with other PHP wrappers for select syscall.
robert dot reichel at boboton dot com
20.10.2007 14:54
I was testing PHP code provided by in curl_multi_exec section of PHP Manual.

The part of the code '$err = curl_error($conn[$i])' should return error message for each cURL session, but it does not.
The function curl_error() works well with the curl_exec(). Is there any other solution for getting session error message with curl_multi_exec() or there is a bug in cURL library.

The script was tested with Windows XP and PHP-5.2.4
shichuanr at gmail dot com
6.09.2007 14:12
For people who have problem running the above example script(Example 425. curl_multi_exec()), you can alter it a little bit to make it work properly by replacing the existing chunk of code with the one below:

// create both cURL resources
$ch1 = curl_init();
$ch2 = curl_init();

// set URL and other appropriate options
curl_setopt($ch1, CURLOPT_URL, "");
curl_setopt($ch1, CURLOPT_HEADER, 0);
curl_setopt($ch2, CURLOPT_URL, "");
curl_setopt($ch2, CURLOPT_HEADER, 0);

//create the multiple cURL handle
$mh = curl_multi_init();

//add the two handles

//execute the handles
do {
} while (
$running > 0);
//close the handles

substr("iscampifriese",1,11) dot " at beer dot com"
4.07.2007 15:10
If you are using mulit handles and you wish to re-execute any of them (if they timed out or something), then you need to remove the handles from the mulit-handle and then re-add them in order to get them to re-execute.  Otherwise cURL will just give you back the same results again without actually retrying.

PHP Powered Diese Seite bei
The PHP manual text and comments are covered by the Creative Commons Attribution 3.0 License © the PHP Documentation Group - Impressum - mail("TO:Reinhard Neidl",...)