X-Git-Url: http://wagnertech.de/git?a=blobdiff_plain;f=SL%2FBackgroundJob%2FCsvImport.pm;h=7df8b9b476e050ff1bba714cf971f3bde6d7a1fd;hb=fba56a49429a170e00a0f26f15784303eb99e108;hp=0300179640f4750c0fc31935f610240d7bfc72e0;hpb=b2321d6af4052de92acf0e3e1e81e29380a4d0f8;p=kivitendo-erp.git diff --git a/SL/BackgroundJob/CsvImport.pm b/SL/BackgroundJob/CsvImport.pm index 030017964..7df8b9b47 100644 --- a/SL/BackgroundJob/CsvImport.pm +++ b/SL/BackgroundJob/CsvImport.pm @@ -5,7 +5,6 @@ use strict; use parent qw(SL::BackgroundJob::Base); use YAML (); -use SL::Controller::CsvImport; use SL::DB::CsvImportProfile; use SL::SessionFile::Random; @@ -51,18 +50,18 @@ sub run { $self->{db_obj} = shift; $self->do_import; - - $self->cleanup; } sub do_import { my ($self) = @_; + require SL::Controller::CsvImport; my $c = SL::Controller::CsvImport->new; my $job = $self->{db_obj}; $c->profile($self->profile); $c->type($job->data_as_hash->{type}); + $c->{employee_id} = $job->data_as_hash->{employee_id}; my $test = $job->data_as_hash->{test}; @@ -76,7 +75,7 @@ sub do_import { plan => { 'parsing csv' => 1, 'building data' => 2, - ( 'saving data' => 3, )x!!$test, + ( 'saving data' => 3, )x!$test, 'building report' => ($test ? 3 : 4), }, num_phases => ($test ? 3 : 4), @@ -84,12 +83,11 @@ sub do_import { $c->add_progress_tracker($self); - $c->test_and_import(test => 1, session_id => $job->data_as_hash->{session_id}); + $c->test_and_import(test => $test, session_id => $job->data_as_hash->{session_id}); if ($c->errors) { $job->set_data( errors => $c->errors, - progress => -1, )->save; } else { @@ -111,10 +109,6 @@ sub track_progress { $self->{db_obj}->save; } -sub cleanup { - -} - 1; __END__