[MCOL-3514] Make cpimport read from data in S3 buckets Created: 2019-09-24 Updated: 2020-02-11 Resolved: 2020-02-11 |
|
| Status: | Closed |
| Project: | MariaDB ColumnStore |
| Component/s: | None |
| Affects Version/s: | None |
| Fix Version/s: | 1.4.0 |
| Type: | New Feature | Priority: | Major |
| Reporter: | Andrew Hutchings (Inactive) | Assignee: | Daniel Lee (Inactive) |
| Resolution: | Fixed | Votes: | 0 |
| Labels: | None | ||
| Sub-Tasks: |
|
|||||||||||||||
| Sprint: | 2019-06, 2020-1, 2020-2 |
| Description |
|
cpimport needs new options to allow it to read a source file from an Amazon S3 bucket. |
| Comments |
| Comment by Andrew Hutchings (Inactive) [ 2019-09-24 ] | |||||
|
Implementation details... New options for cpimport:
The hostname only needs to be supplied if the S3 server is not Amazon's. It will then use the path/filename to retrieve the file from the S3 bucket into memory and apply it. You will need enough RAM spare to take the entire CSV file. | |||||
| Comment by Daniel Lee (Inactive) [ 2019-09-27 ] | |||||
|
Build tested: 1.4.0-1 [dlee@master centos7]$ cat gitversionInfo.txt Running test on multi-node (1um2pm) returned an error /usr/local/mariadb/columnstore/bin/cpimport mytest lineitem lineitem.tbl -y [mykey] -K [mysecret] -t dleeqatest -g us-west-2 2019-09-27 18:25:08 (9124) ERR : Could not open Input file lineitem.tbl It worked in single node stack: /usr/local/mariadb/columnstore/bin/cpimport mytest lineitem lineitem.tbl -y [mykey] -K [mysecret] -t dleeqatest -g us-west-2 Using table OID 3017 as the default JOB ID [root@localhost ~]# mcsmysql mytest Welcome to the MariaDB monitor. Commands end with ; or \g. Copyright (c) 2000, 2018, Oracle, MariaDB Corporation Ab and others. Type 'help;' or '\h' for help. Type '\c' to clear the current input statement. MariaDB [mytest]> select count
----------
---------- | |||||
| Comment by Daniel Lee (Inactive) [ 2020-02-11 ] | |||||
|
Verified sub-tasks. Closing this ticket now. |