You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* Changed location of installation of GCP cli
* Adding packaging tests
* Test Unit Tests
* Test Functional Tests
* Fix for downloader
* Adding an specific version to not have issues on instalation
* Interconnection experiment
* Test updated workflow
* Fixed out_report.xml path
* Interconnection for GCP
* Adding updates for wheel and setuptools
* Experiment requirements
* Changes on interconnection proyect and setup
* Changes to ignore copied files
* Adding fix to handle encoding issues
* Adding changes to readme
* Taking current folder to get packages
* Fix typo
Signed-off-by: Felipe Leza Alvarez <[email protected]>
* Inter connection sample
* Hot fix for python 3.8
* Fix for setup code
* ignoring sample file
* Renaming correctly interconnection
* Fix on upgrade for pip, wheel and setuptools
* Interoperability POC sample fix
* Fix for .sample copy
* Setup for bash
* Hotfix: change "test" strings for "test_unittest" so that unit tests won't use files used by samples.
* Fixing requirements and setup
* Removing wheel update from setup
* Fixes the problem to upload dataset to GCP.
* Fixed unwanted changes on main branch
* Fixed aws functional test name
* Upload a folder using the aws connector
* Last changes on Interoperability are applied
* Updated Licence
* Removinb code of conduct reference, we have not
* Skips row 0 from excel when creating dataframe.
* Refactor
* delete unused gitignore
* change access keys format
* Refactored names
* first version license header
* Remoivng readme files on this branch
* Deleting files created from setup
* headers
* Adding headers into main packages
* Adding headers on sample code
* Updating files for publishing
* Get GCP credentials
* Changed setup.sh file
* Removed init file on interoperability folder
* Big refactoring, moving data_connector into datasets
* Complete sample link
* Test WF
* Fixed path for unit tests
* Changed trigger to PR
* Create sample link
* Merging readmes
* Removing licence only for data connector
* Missed recursive flag
* delete unused names
* Ignore outputs of jupyter notebook
* Removing commented block
* Removing deprecated folder
* Include headers in init files
* Adding headers in all tests
* Removing coverage omit from tox configuration file (This only works on MZ env)
* Removed gcp auth commands instead of commenting them
* Adding headers
* Removed sensitive information on gcp
* Removing values to make it more easy for usersfill it
* Solving names to public repo
* Removing default values
* Removing sample values
* Removing values
* Fixing path on script
* Removing extra files
* add security file
* Updating files and structure for publishing
* Updates for packaging
* Updating readme
* Updating metadata
* Updating source code
* Updating git ignore
* Updating git ignore
* updating metadata
* Updating gitgitnore
* removing extracted files
* Updating repo
* removing dataset egg info
* Updating file permissions
* Updating file permissions
* removing key
* Updating imports
* Updating reame
* Removing error
* removing data_connector changes
* Updating conda recipes
* Fixed bugs in setup.sh
Added azure src and dependencies.
* Removing conda folders
* Updating blank space at the end of files
* Updating readme
* Validation/scans (#56)
* Fixed dataset_api requirements file
* Merging from data_connector
* Updating gitignore
* Returning depencencies
* Returning training code
* Creating and re naming sample files
* Adding format
* New readme proposals
* Fix on toml to avoid refactor
* Readme agenda
* Conda folder is unevitable
* Exclud conda and egg folders
* Adding badages in main readme... will see if we should use rst format for main readme only
* Simple entry point for sample doc
* Change header for sub_linked section
* Modifications to current lass invocation
* Adding relative link to documentation in AWS main readme file
* Terms and conditions requirements update
* Changes on Azure Readmi file
* Removing previous terms and conditions
* Updating path for datasets_urls
* Updating path for datasets_urls
* Removing data connector changes
* Updating blank last line
* Updated documentation with curren code functionality
* Update documentation
* Added code sample for upload, download and list blobs for oauth
* first definition on dcp readme for bigquery
* Sample connection with oauth
* Adding readme sample for gcp service account connection with GCP
* Connection documentation finished
* Updating TPP file
* updating with feedback
---------
Signed-off-by: gera-aldama <[email protected]>
Signed-off-by: Felipe Leza Alvarez <[email protected]>
Co-authored-by: Miguel Pineda <[email protected]>
Co-authored-by: Gerardo Dominguez <[email protected]>
Co-authored-by: gera-aldama <[email protected]>
Co-authored-by: Felipe Leza Alvarez <[email protected]>
Co-authored-by: aagalleg <[email protected]>
Co-authored-by: Leza Alvarez, Felipe <[email protected]>
Co-authored-by: ma-pineda <[email protected]>
data_connector is a tool to connect to AzureML, Azure blob, GCP storage, GCP Big Query and AWS storage S3.
4
16
The goal is provide all cloud managers in one place and provide documentation for an easy integration.
5
17
6
-
For more details, visit the [Data Connector](repo link) GitHub repository.
18
+
For more details, visit the [Data Connector](https://github.com/IntelAI/models/tree/master/datasets/data_connector) GitHub repository.
19
+
<br/><br/>
7
20
8
21
## Hardware Requirements
9
22
---
10
23
The hardware should comply with the same requirements that the cloud service.
24
+
<br/><br/>
11
25
12
26
## How it Works
13
27
---
@@ -20,7 +34,7 @@ The package contains the following modules:
20
34
| data_connector.azure |
21
35
22
36
Each module is capable of connect, download and upload operation to it-s corresponding cloud service.
23
-
37
+
<br/><br/>
24
38
25
39
## Getting Started with data_connector
26
40
---
@@ -33,23 +47,15 @@ conda activate venv
33
47
34
48
You can install the package with:
35
49
```bash
36
-
python -m pip install intel-cloud-data-connector
50
+
python -m pip install cloud-data-connector
37
51
```
38
52
39
-
Please follow module specific documentation for use case, hands-examples. This documentation can be found inside the package.
53
+
Please follow module specific documentation for use case, hands-examples.
40
54
1. data_connector/azure/README.md
41
55
2. data_connector/azure/AzureML.md
42
56
3. data_connector/aws/README.md
43
57
4. data_connector/gcp/README.md
44
-
45
-
<!---
46
-
## Learn More
47
-
---
48
-
For more information about data_connector, see these guides and software resources:
49
-
- github/repo/link
50
-
TODO: Update public repo
51
-
-->
52
-
58
+
<br/><br/>
53
59
54
60
## Getting Started with data_connector.azure
55
61
---
@@ -102,11 +108,7 @@ Data connector provides a tool to connect to Azure ML workspaces and upload conf
102
108
```
103
109
How to get a [Connection String](https://learn.microsoft.com/en-us/answers/questions/1071173/where-can-i-find-storage-account-connection-string)?
Data Connector for AWS S3 allows you to connect to S3 buckets and list contents, download and upload files.
4
4
@@ -8,7 +8,8 @@ To access S3 buckets, you will need to sign up for an AWS account and create acc
8
8
9
9
Access keys consist of an access key ID and secret access key, which are used to sign programmatic requests that you make to AWS.
10
10
11
-
## Hot to get your access key ID and secret access key
11
+
How to get your access key ID and secret access key
12
+
---
12
13
13
14
1. Open the IAM console at https://console.aws.amazon.com/iam/.
14
15
2. On the navigation menu, choose Users.
@@ -34,31 +35,50 @@ You can add more configuration settings listed [here](https://boto3.amazonaws.co
34
35
You need to import the DataConnector class.
35
36
36
37
```python
37
-
from data_connector.aws.connectorimport Connector
38
+
from data_connector.aws import Connectoras aws_connector
38
39
```
39
40
40
41
Connector class has the method connect(), it creates an AWS S3 object, by default the function will create a S3 connector using the credentials saved in your environment variables.
41
42
42
43
```python
43
-
connector=Connector()
44
+
aws_bucket_connector=aws_connector()
44
45
```
45
46
46
47
Call the connect() method, this will return a connection object for S3.
You can import an Uploader class and use the upload method to send a file from you local machine to a bucket. You need to add the connector object to Uploader constructor.
93
104
94
105
```python
95
-
from data_connector.aws.uploaderimport Uploader
96
-
from data_connector.aws.connector import Connector
106
+
from data_connector.aws import Uploaderas aws_uploader
107
+
from data_connector.awsimport Connectoras aws_connector
0 commit comments