# |
Sep 13th 2019, 16:24 |
challgren |
Plus you can see whats been cleaned and what failed cleanup |
# |
Sep 13th 2019, 16:23 |
louis |
I’m going to try this way ! I agree, never thought about it :expressionless: Thank you @challgren |
# |
Sep 13th 2019, 16:22 |
challgren |
A little bit cleaner IMHO |
# |
Sep 13th 2019, 16:22 |
challgren |
Yeah same idea except the queue handles it |
# |
Sep 13th 2019, 16:21 |
louis |
Okay, in previous projects I was writing CakePHP shell planned with a cron task. this is why I’m looking for better solution :slightly_smiling_face: |
# |
Sep 13th 2019, 16:20 |
challgren |
That way if the user wants to redownload it again, Im not hammering the server with another pdf creation/zip process and can quickly serve them the “cached” package |
# |
Sep 13th 2019, 16:19 |
challgren |
Using https://github.com/dereuromark/cakephp-queue |
# |
Sep 13th 2019, 16:19 |
challgren |
But what I do is create my pdfs/zips and then create a queue task to clean it up in an hour |
# |
Sep 13th 2019, 16:18 |
challgren |
Nah Im just waking up so the mt dew hasnt kicked in yet |
# |
Sep 13th 2019, 16:18 |
louis |
Should I explain in another way ? |
# |
Sep 13th 2019, 16:18 |
challgren |
Oh probably do a queue task |
# |
Sep 13th 2019, 16:18 |
louis |
@challgren Yeap I’m using it ;) . Sorry if my question is not well formulated |
# |
Sep 13th 2019, 16:17 |
challgren |
@louis https://github.com/FriendsOfCake/CakePdf is awesome! |
# |
Sep 13th 2019, 16:16 |
louis |
Hello everyone, I’m looking for advices for PDF generation. At the moment i’ve a method that saving multiple Pdf files into TMP directory (named with a uuid) adnd then I create an archive of theses Pdf. The problem is what is a good way to remove this directory after serving file to user using `response->withFile()` |
# |
Sep 13th 2019, 16:15 |
ruano84 |
Hi guys |
# |
Sep 13th 2019, 14:35 |
admad |
welcome |
# |
Sep 13th 2019, 14:33 |
felipe.marinho |
Nice, thank you :pray: |
# |
Sep 13th 2019, 14:29 |
admad |
https://dev.mysql.com/doc/refman/8.0/en/memory-storage-engine.html |
# |
Sep 13th 2019, 14:29 |
admad |
*storage engine |
# |
Sep 13th 2019, 14:28 |
admad |
also if possible use 'MEMORY' engine for the temp table |
# |
Sep 13th 2019, 14:27 |
felipe.marinho |
Awesome... a temp table and an action to match those 2 tables... I'll try this too... Thank you @admad |
# |
Sep 13th 2019, 14:26 |
admad |
*if your |
# |
Sep 13th 2019, 14:26 |
admad |
then do whatever processing you need |
# |
Sep 13th 2019, 14:25 |
admad |
if you table columns don't directly match csv columns then first dump the csv in a temp table with matching columns |
# |
Sep 13th 2019, 14:24 |
felipe.marinho |
the problem is to match the fields from the CSV and the table columns... |
# |
Sep 13th 2019, 14:23 |
admad |
you should just upload file and use a shell / command to do the importing in background |
# |
Sep 13th 2019, 14:23 |
admad |
ideally importing 500,000 lines of csv should be even done in a web request. |
# |
Sep 13th 2019, 14:22 |
felipe.marinho |
Yes, his approach was to use the input type="hidden"... I'll try to upload the CSV, match the fields, and after use the uploaded CSV. |
# |
Sep 13th 2019, 14:22 |
admad |
sorry but i am not going to bother reading that :slightly_smiling_face: |
# |
Sep 13th 2019, 14:21 |
felipe.marinho |
I adapted this tutorial: https://quickadminpanel.com/blog/how-to-import-csv-in-laravel-and-choose-matching-fields/ |
# |
Sep 13th 2019, 14:20 |
admad |
if not possible then query builder to do multi insert |
# |
Sep 13th 2019, 14:20 |
admad |
next don't use ORM methods to insert. if server allows Best would be to `LOAD DATA INFILE` syntax to directly make mysql use the csv file to insert rows |
# |
Sep 13th 2019, 14:18 |
admad |
also make sure post_max_size and max_file_size are set to large enough values in php.ini to allow large file uploads |
# |
Sep 13th 2019, 14:18 |
admad |
using hidden input to pass csv data sounds like a very brittle approach. use a file input to upload |
# |
Sep 13th 2019, 14:17 |
ckjksl |
Locally will be faster, but are you trying to save 500,000 entities all at once? |
# |
Sep 13th 2019, 14:15 |
slackebot |
approach is the best option for now... locally it will be faster. |
# |
Sep 13th 2019, 14:15 |
felipe.marinho |
I'm using MYSQL 5.5, I read about update the mysql version, I'll try to install the MariaDB updated and test it... and maybe latter try to create all this entities using an uploaded csv... I tried to use the file inside the TMP folder, but after it uploads and use it inside the action the server delete the file... I saved the tmp file path in the input type="hidden" i created first, but the file was deleted :( I think the upload |
# |
Sep 13th 2019, 14:11 |
ckjksl |
Maybe instead of trying to create them all at once, you can stagger it so it saves 100 (or whatever number you want) at a time? 1000 seconds isn't very long |
# |
Sep 13th 2019, 14:07 |
felipe.marinho |
@ckjksl The PHP Timeout in the php.ini is max_execution_time=1000, and it crashes instantaneously, not even a single second lol... Yes, I'm using saveMany to save all those data... |
# |
Sep 13th 2019, 14:02 |
slackebot |
=> $this->toIso8601String(), 'timezone' => $this->getTimezone()->getName(), 'fixedNowTime' => static::hasTestNow() ? static::getTestNow()->toIso8601String() : false ]; ``` seems to trigger it. |
# |
Sep 13th 2019, 14:02 |
sdevore |
I realize I asked this yesterday afternoon but any thoughts on how to find where xdebug is crashing in __debugInfo() after updating from 3.6 to 3.7, I updated chronos to 1.2.8 since that seemed to mention some issues with __debugInfo() the line in my code that seems to trigger it is $enrolled_course->set(‘course’, $overlap); and stepping though it is almost seems that something in DateFormat::__debugInfo(); ``` return [ 'ti |