MCE::Child(3pm) | User Contributed Perl Documentation | MCE::Child(3pm) |
MCE::Child - A threads-like parallelization module compatible with Perl 5.8
This document describes MCE::Child version 1.889
use MCE::Child; MCE::Child->init( max_workers => 'auto', # default undef, unlimited # Specify a percentage. MCE::Child 1.876+. max_workers => '25%', # 4 on HW with 16 lcores max_workers => '50%', # 8 on HW with 16 lcores child_timeout => 20, # default undef, no timeout posix_exit => 1, # default undef, CORE::exit void_context => 1, # default undef on_start => sub { my ( $pid, $ident ) = @_; ... }, on_finish => sub { my ( $pid, $exit, $ident, $signal, $error, @ret ) = @_; ... } ); MCE::Child->create( sub { print "Hello from child\n" } )->join(); sub parallel { my ($arg1) = @_; print "Hello again, $arg1\n" if defined($arg1); print "Hello again, $_\n"; # same thing } MCE::Child->create( \¶llel, $_ ) for 1 .. 3; my @procs = MCE::Child->list(); my @pids = MCE::Child->list_pids(); my @running = MCE::Child->list_running(); my @joinable = MCE::Child->list_joinable(); my @count = MCE::Child->pending(); # Joining is orderly, e.g. child1 is joined first, child2, child3. $_->join() for @procs; # (or) $_->join() for @joinable; # Joining occurs immediately as child processes complete execution. 1 while MCE::Child->wait_one(); my $child = mce_child { foreach (@files) { ... } }; $child->join(); if ( my $err = $child->error() ) { warn "Child error: $err\n"; } # Get a child's object $child = MCE::Child->self(); # Get a child's ID $pid = MCE::Child->pid(); # $$ $pid = $child->pid(); $pid = MCE::Child->tid(); # tid is an alias for pid $pid = $child->tid(); # Test child objects if ( $child1 == $child2 ) { ... } # Give other workers a chance to run MCE::Child->yield(); MCE::Child->yield(0.05); # Return context, wantarray aware my ($value1, $value2) = $child->join(); my $value = $child->join(); # Check child's state if ( $child->is_running() ) { sleep 1; } if ( $child->is_joinable() ) { $child->join(); } # Send a signal to a child $child->kill('SIGUSR1'); # Exit a child MCE::Child->exit(0); MCE::Child->exit(0, @ret);
MCE::Child is a fork of MCE::Hobo for compatibility with Perl 5.8.
A child is a migratory worker inside the machine that carries the asynchronous gene. Child processes are equipped with "threads"-like capability for running code asynchronously. Unlike threads, each child is a unique process to the underlying OS. The IPC is handled via "MCE::Channel", which runs on all the major platforms including Cygwin and Strawberry Perl.
"MCE::Child" may be used as a standalone or together with "MCE" including running alongside "threads".
use MCE::Child; use MCE::Shared; # synopsis: head -20 file.txt | perl script.pl my $ifh = MCE::Shared->handle( "<", \*STDIN ); # shared my $ofh = MCE::Shared->handle( ">", \*STDOUT ); my $ary = MCE::Shared->array(); sub parallel_task { my ( $id ) = @_; while ( <$ifh> ) { printf {$ofh} "[ %4d ] %s", $., $_; # $ary->[ $. - 1 ] = "[ ID $id ] read line $.\n" ); # dereferencing $ary->set( $. - 1, "[ ID $id ] read line $.\n" ); # faster via OO } } my $child1 = MCE::Child->new( "parallel_task", 1 ); my $child2 = MCE::Child->new( \¶llel_task, 2 ); my $child3 = MCE::Child->new( sub { parallel_task(3) } ); $_->join for MCE::Child->list(); # ditto: MCE::Child->wait_all(); # search array (total one round-trip via IPC) my @vals = $ary->vals( "val =~ / ID 2 /" ); print {*STDERR} join("", @vals);
FUNCTION may either be the name of a function, an anonymous subroutine, or a code ref.
my $child = MCE::Child->create( "func_name", ... ); # or my $child = MCE::Child->create( sub { ... }, ... ); # or my $child = MCE::Child->create( \&func, ... );
The "ident" option is used by callback functions "on_start" and "on_finish" for identifying the started and finished child process respectively.
my $child1 = MCE::Child->create( { posix_exit => 1 }, sub { ... } ); $child1->join; my $child2 = MCE::Child->create( { child_timeout => 3 }, sub { sleep 1 for ( 1 .. 9 ); } ); $child2->join; if ( $child2->error() eq "Child timed out\n" ) { ... }
The "new()" method is an alias for "create()".
my $child = mce_child { foreach (@files) { ... } }; $child->join(); if ( my $err = $child->error() ) { warn("Child error: $err\n"); }
The context (void, scalar or list) for the return value(s) for "join" is determined at the time of joining and mostly "wantarray" aware.
my $child1 = MCE::Child->create( sub { my @res = qw(foo bar baz); return (@res); }); my @res1 = $child1->join(); # ( foo, bar, baz ) my $res1 = $child1->join(); # baz my $child2 = MCE::Child->create( sub { return 'foo'; }); my @res2 = $child2->join(); # ( foo ) my $res2 = $child2->join(); # foo
if ( $child1 == $child2 ) { print("Child objects are the same\n"); } # or if ( $child1 != $child2 ) { print("Child objects differ\n"); }
$child->exit()->join(); ... $child->join(); # later
MCE::Child->create( 'task1', $_ ) for 1 .. 4; $_->join for MCE::Child->list(); MCE::Child->create( 'task2', $_ ) for 1 .. 4; $_->join for MCE::Child->list(); MCE::Child->create( 'task3', $_ ) for 1 .. 4; $_->join for MCE::Child->list(); MCE::Child->finish();
MCE::Child->init( max_workers => 'auto', # default undef, unlimited # Specify a percentage. MCE::Child 1.876+. max_workers => '25%', # 4 on HW with 16 lcores max_workers => '50%', # 8 on HW with 16 lcores child_timeout => 20, # default undef, no timeout posix_exit => 1, # default undef, CORE::exit void_context => 1, # default undef on_start => sub { my ( $pid, $ident ) = @_; ... }, on_finish => sub { my ( $pid, $exit, $ident, $signal, $error, @ret ) = @_; ... } ); # Identification given as an option or the 1st argument. for my $key ( 'aa' .. 'zz' ) { MCE::Child->create( { ident => $key }, sub { ... } ); MCE::Child->create( $key, sub { ... } ); } MCE::Child->wait_all;
Set "max_workers" if you want to limit the number of workers by waiting automatically for an available slot. Specify a percentage or "auto" to obtain the number of logical cores via "MCE::Util::get_ncpu()".
Set "child_timeout", in number of seconds, if you want the child process to terminate after some time. The default is 0 for no timeout.
Set "posix_exit" to avoid all END and destructor processing. Constructing MCE::Child inside a thread implies 1 or if present CGI, FCGI, Coro, Curses, Gearman::Util, Gearman::XS, LWP::UserAgent, Mojo::IOLoop, STFL, Tk, Wx, or Win32::GUI.
Set "void_context" to create the child process in void context for the return value. Otherwise, the return context is wantarray-aware for "join()" and "result()" and determined when retrieving the data.
The callback options "on_start" and "on_finish" are called in the parent process after starting the worker and later when terminated. The arguments for the subroutines were inspired by Parallel::ForkManager.
The parameters for "on_start" are the following:
- pid of the child process - identification (ident option or 1st arg to create)
The parameters for "on_finish" are the following:
- pid of the child process - program exit code - identification (ident option or 1st arg to create) - exit signal id - error message from eval inside MCE::Child - returned data
$child->kill('SIG...')->join();
The following is a parallel demonstration comparing "MCE::Shared" against "Redis" and "Redis::Fast" on a Fedora 23 VM. Joining begins after all workers have been notified to quit.
use Time::HiRes qw(time); use Redis; use Redis::Fast; use MCE::Child; use MCE::Shared; my $redis = Redis->new(); my $rfast = Redis::Fast->new(); my $array = MCE::Shared->array(); sub parallel_redis { my ($_redis) = @_; my ($count, $quit, $len) = (0, 0); # instead, use a flag to exit loop $SIG{'QUIT'} = sub { $quit = 1 }; while () { $len = $_redis->rpush('list', $count++); last if $quit; } $count; } sub parallel_array { my ($count, $quit, $len) = (0, 0); # do not exit from inside handler $SIG{'QUIT'} = sub { $quit = 1 }; while () { $len = $array->push($count++); last if $quit; } $count; } sub benchmark_this { my ($desc, $num_procs, $timeout, $code, @args) = @_; my ($start, $total) = (time(), 0); MCE::Child->new($code, @args) for 1..$num_procs; sleep $timeout; # joining is not immediate; ok $_->kill('QUIT') for MCE::Child->list(); # joining later; ok $total += $_->join() for MCE::Child->list(); printf "$desc <> duration: %0.03f secs, count: $total\n", time() - $start; sleep 0.2; } benchmark_this('Redis ', 8, 5.0, \¶llel_redis, $redis); benchmark_this('Redis::Fast', 8, 5.0, \¶llel_redis, $rfast); benchmark_this('MCE::Shared', 8, 5.0, \¶llel_array);
@procs = MCE::Child->list();
@pids = MCE::Child->list_pids(); $SIG{INT} = $SIG{HUP} = $SIG{TERM} = sub { # Signal workers all at once CORE::kill('KILL', MCE::Child->list_pids()); exec('reset'); };
@procs = MCE::Child->list_running();
@procs = MCE::Child->list_joinable();
$count = MCE::Child->pending();
use MCE::Child; use Time::HiRes qw(sleep); sub task { my ($id) = @_; sleep $id * 0.333; return $id; } MCE::Child->create('task', $_) for ( reverse 1 .. 3 ); # 1 while MCE::Child->wait_one(); while ( my $child = MCE::Child->wait_one() ) { my $err = $child->error() || 'no error'; my $res = $child->result(); my $pid = $child->pid(); print "[$pid] $err : $res\n"; }
Like "join" described above, the context (void, scalar or list) for the return value(s) is determined at the time "result" is called and mostly "wantarray" aware.
my $child1 = MCE::Child->create( sub { my @res = qw(foo bar baz); return (@res); }); my @res1 = $child1->result(); # ( foo, bar, baz ) my $res1 = $child1->result(); # baz my $child2 = MCE::Child->create( sub { return 'foo'; }); my @res2 = $child2->result(); # ( foo ) my $res2 = $child2->result(); # foo
pid: $$ process id tid: $$ alias for pid
pid: $$ process id tid: $$ alias for pid
The "waitone" and "waitall" methods are aliases for compatibility with "MCE::Hobo".
use MCE::Child; use Time::HiRes qw(sleep); sub task { my $id = shift; sleep $id * 0.333; return $id; } MCE::Child->create('task', $_) for ( reverse 1 .. 3 ); # join, traditional use case $_->join() for MCE::Child->list(); # wait_one, simplistic use case 1 while MCE::Child->wait_one(); # wait_one while ( my $child = MCE::Child->wait_one() ) { my $err = $child->error() || 'no error'; my $res = $child->result(); my $pid = $child->pid(); print "[$pid] $err : $res\n"; } # wait_all my @procs = MCE::Child->wait_all(); for ( @procs ) { my $err = $_->error() || 'no error'; my $res = $_->result(); my $pid = $_->pid(); print "[$pid] $err : $res\n"; }
The default "floating_seconds" is 0.008 and 0.015 on UNIX and Windows, respectively. Pass 0 if simply wanting to give other workers a chance to run.
# total run time: 1.00 second MCE::Child->create( sub { MCE::Child->yield(0.25) } ) for 1 .. 4; MCE::Child->wait_all();
Threads-like detach capability was added starting with the 1.867 release.
A threads example is shown first followed by the MCE::Child example. All one needs to do is set the CHLD signal handler to IGNORE. Unfortunately, this works on UNIX platforms only. The child process restores the CHLD handler to default, so is able to deeply spin workers and reap if desired.
use threads; for ( 1 .. 8 ) { async { # do something }->detach; } use MCE::Child; # Have the OS reap workers automatically when exiting. # The on_finish option is ignored if specified (no-op). # Ensure not inside a thread on UNIX platforms. $SIG{CHLD} = 'IGNORE'; for ( 1 .. 8 ) { mce_child { # do something }; } # Optionally, wait for any remaining workers before leaving. # This is necessary if workers are consuming shared objects, # constructed via MCE::Shared. MCE::Child->wait_all;
The following is another way and works on Windows. Here, the on_finish handler works as usual.
use MCE::Child; MCE::Child->init( on_finish = sub { ... }, ); for ( 1 .. 8 ) { $_->join for MCE::Child->list_joinable; mce_child { # do something }; } MCE::Child->wait_all;
MCE::Child behaves similarly to threads for the most part. It also provides Parallel::ForkManager-like capabilities. The "Parallel::ForkManager" example is shown first followed by a version using "MCE::Child".
use strict; use warnings; use Parallel::ForkManager; use Time::HiRes 'time'; my $start = time; my $pm = Parallel::ForkManager->new(10); $pm->set_waitpid_blocking_sleep(0); $pm->run_on_finish( sub { my ($pid, $exit_code, $ident, $exit_signal, $core_dumped, $resp) = @_; print "child $pid completed: $ident => ", $resp->[0], "\n"; }); DATA_LOOP: foreach my $data ( 1..2000 ) { # forks and returns the pid for the child my $pid = $pm->start($data) and next DATA_LOOP; my $ret = [ $data * 2 ]; $pm->finish(0, $ret); } $pm->wait_all_children; printf STDERR "duration: %0.03f seconds\n", time - $start;
use strict; use warnings; use MCE::Child 1.843; use Time::HiRes 'time'; my $start = time; MCE::Child->init( max_workers => 10, on_finish => sub { my ($pid, $exit_code, $ident, $exit_signal, $error, $resp) = @_; print "child $pid completed: $ident => ", $resp->[0], "\n"; } ); foreach my $data ( 1..2000 ) { MCE::Child->create( $data, sub { [ $data * 2 ]; }); } MCE::Child->wait_all; printf STDERR "duration: %0.03f seconds\n", time - $start;
MCE::Hobo uses MCE::Shared to retrieve data during reaping. MCE::Child uses MCE::Channel, no shared-manager. Version Cygwin Windows Linux macOS FreeBSD MCE::Child 1.843 19.099s 17.091s 0.965s 1.534s 1.229s MCE::Hobo 1.843 20.514s 19.594s 1.246s 1.629s 1.613s P::FM 1.20 19.703s 19.235s 0.875s 1.445s 1.346s MCE::Child 1.843 20.426s 18.417s 1.116s 1.632s 1.338s Moo loaded MCE::Hobo 1.843 21.809s 20.810s 1.407s 1.759s 1.722s Moo loaded P::FM 2.02 21.668s 25.927s 1.882s 2.612s 2.483s Moo used
MCE::Child->init( posix_exit => 1, ... ); MCE::Hobo->init( posix_exit => 1, ... ); Version Cygwin Windows Linux macOS FreeBSD MCE::Child 1.843 19.815s ignored 0.824s 1.284s 1.245s Moo loaded MCE::Hobo 1.843 21.029s ignored 0.953s 1.335s 1.439s Moo loaded
This demonstration constructs two queues, two handles, starts the shared-manager process if needed, and spawns four workers. For this demonstration, am chunking 64 URLs per job. In reality, one may run with 200 workers and chunk 300 URLs on a 24-way box.
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ # perl demo.pl -- all output # perl demo.pl >/dev/null -- mngr/child output # perl demo.pl 2>/dev/null -- show results only # # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ use strict; use warnings; use AnyEvent; use AnyEvent::HTTP; use Time::HiRes qw( time ); use MCE::Child; use MCE::Shared; # Construct two queues, input and return. my $que = MCE::Shared->queue(); my $ret = MCE::Shared->queue(); # Construct shared handles for serializing output from many workers # writing simultaneously. This prevents garbled output. mce_open my $OUT, ">>", \*STDOUT or die "open error: $!"; mce_open my $ERR, ">>", \*STDERR or die "open error: $!"; # Spawn workers early for minimum memory consumption. MCE::Child->create({ posix_exit => 1 }, 'task', $_) for 1 .. 4; # Obtain or generate input data for workers to process. my ( $count, @urls ) = ( 0 ); push @urls, map { "http://127.0.0.$_/" } 1..254; push @urls, map { "http://192.168.0.$_/" } 1..254; # 508 URLs total while ( @urls ) { my @chunk = splice(@urls, 0, 64); $que->enqueue( { ID => ++$count, INPUT => \@chunk } ); } # So that workers leave the loop after consuming the queue. $que->end(); # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ # Loop for the manager process. The manager may do other work if # need be and periodically check $ret->pending() not shown here. # # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ my $start = time; printf {$ERR} "Mngr - entering loop\n"; while ( $count ) { my ( $result, $failed ) = $ret->dequeue( 2 ); # Remove ID from result, so not treated as a URL item. printf {$ERR} "Mngr - received job %s\n", delete $result->{ID}; # Display the URL and the size captured. foreach my $url ( keys %{ $result } ) { printf {$OUT} "%s: %d\n", $url, length($result->{$url}) if $result->{$url}; # url has content } # Display URLs could not reach. if ( @{ $failed } ) { foreach my $url ( @{ $failed } ) { print {$OUT} "Failed: $url\n"; } } # Decrement the count. $count--; } MCE::Child->wait_all(); printf {$ERR} "Mngr - exiting loop\n\n"; printf {$ERR} "Duration: %0.3f seconds\n\n", time - $start; exit; # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ # Child processes enqueue two items ( $result and $failed ) per each # job for the manager process. Likewise, the manager process dequeues # two items above. Optionally, child processes may include the ID in # the result. # # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ sub task { my ( $id ) = @_; printf {$ERR} "Child $id entering loop\n"; while ( my $job = $que->dequeue() ) { my ( $result, $failed ) = ( { ID => $job->{ID} }, [ ] ); # Walk URLs, provide a hash and array refs for data. printf {$ERR} "Child $id running job $job->{ID}\n"; walk( $job, $result, $failed ); # Send results to the manager process. $ret->enqueue( $result, $failed ); } printf {$ERR} "Child $id exiting loop\n"; } sub walk { my ( $job, $result, $failed ) = @_; # Yielding is critical when running an event loop in parallel. # Not doing so means that the app may reach contention points # with the firewall and likely impose unnecessary hardship at # the OS level. The idea here is not to have multiple workers # initiate HTTP requests to a batch of URLs at the same time. # Yielding behaves similarly like scatter to have the child # process run solo for a fraction of time. MCE::Child->yield( 0.03 ); my $cv = AnyEvent->condvar(); # Populate the hash ref for the URLs it could reach. # Do not mix AnyEvent timeout with child timeout. # Therefore, choose event timeout when available. foreach my $url ( @{ $job->{INPUT} } ) { $cv->begin(); http_get $url, timeout => 2, sub { my ( $data, $headers ) = @_; $result->{$url} = $data; $cv->end(); }; } $cv->recv(); # Populate the array ref for URLs it could not reach. foreach my $url ( @{ $job->{INPUT} } ) { push @{ $failed }, $url unless (exists $result->{ $url }); } return; } __END__ $ perl demo.pl Child 1 entering loop Child 2 entering loop Child 3 entering loop Mngr - entering loop Child 2 running job 2 Child 3 running job 3 Child 1 running job 1 Child 4 entering loop Child 4 running job 4 Child 2 running job 5 Mngr - received job 2 Child 3 running job 6 Mngr - received job 3 Child 1 running job 7 Mngr - received job 1 Child 4 running job 8 Mngr - received job 4 http://192.168.0.1/: 3729 Child 2 exiting loop Mngr - received job 5 Child 3 exiting loop Mngr - received job 6 Child 1 exiting loop Mngr - received job 7 Child 4 exiting loop Mngr - received job 8 Mngr - exiting loop Duration: 4.131 seconds
Making an executable is possible with the PAR::Packer module. On the Windows platform, threads, threads::shared, and exiting via threads are necessary for the binary to exit successfully.
# https://metacpan.org/pod/PAR::Packer # https://metacpan.org/pod/pp # # pp -o demo.exe demo.pl # ./demo.exe use strict; use warnings; use if $^O eq "MSWin32", "threads"; use if $^O eq "MSWin32", "threads::shared"; # Include minimum dependencies for MCE::Child. # Add other modules required by your application here. use Storable (); use Time::HiRes (); # use IO::FDPass (); # optional: for condvar, handle, queue # use Sereal (); # optional: for faster serialization use MCE::Child; use MCE::Shared; # For PAR to work on the Windows platform, one must include manually # any shared modules used by the application. # use MCE::Shared::Array; # if using MCE::Shared->array # use MCE::Shared::Cache; # if using MCE::Shared->cache # use MCE::Shared::Condvar; # if using MCE::Shared->condvar # use MCE::Shared::Handle; # if using MCE::Shared->handle, mce_open # use MCE::Shared::Hash; # if using MCE::Shared->hash # use MCE::Shared::Minidb; # if using MCE::Shared->minidb # use MCE::Shared::Ordhash; # if using MCE::Shared->ordhash # use MCE::Shared::Queue; # if using MCE::Shared->queue # use MCE::Shared::Scalar; # if using MCE::Shared->scalar # Et cetera. Only load modules needed for your application. use MCE::Shared::Sequence; # if using MCE::Shared->sequence my $seq = MCE::Shared->sequence( 1, 9 ); sub task { my ( $id ) = @_; while ( defined ( my $num = $seq->next() ) ) { print "$id: $num\n"; sleep 1; } } sub main { MCE::Child->new( \&task, $_ ) for 1 .. 3; MCE::Child->wait_all(); } # Main must run inside a thread on the Windows platform or workers # will fail duing exiting, causing the exe to crash. The reason is # that PAR or a dependency isn't multi-process safe. ( $^O eq "MSWin32" ) ? threads->create(\&main)->join() : main(); threads->exit(0) if $INC{"threads.pm"};
MCE::Child emits an error when "is_joinable", "is_running", and "join" isn't called by the managed process, where the child was spawned. This is a limitation in MCE::Child only due to not involving a shared-manager process for IPC.
This use-case is not typical.
The inspiration for "MCE::Child" comes from wanting "threads"-like behavior for processes compatible with Perl 5.8. Both can run side-by-side including safe-use by MCE workers. Likewise, the documentation resembles "threads".
The inspiration for "wait_all" and "wait_one" comes from the "Parallel::WorkUnit" module.
MCE, MCE::Channel, MCE::Shared
Mario E. Roy, <marioeroy AT gmail DOT com>
2023-09-29 | perl v5.36.0 |