How to upload a folder on SageMaker Notebooks - amazon-web-services

I am trying to upload a folder on SageMaker. However, I cannot select any folders, I need to go through the files and upload them one by one. Is there any way to upload the whole directory from local computer to SageMaker?

zip the folder in local, upload it as a zip file and unzip it in Sagemaker terminal

Related

Migrate PyCharm project files to AWS S3

I have a PyCharm project that is stored on my local disk. The project scripts download a file folder that is quite large. I have run out of space on the hard disk and am looking for ways to migrate the data files or download them to an aws s3 bucket I have created.
my current directory:
/Users/$User/PycharmProjects/Project
s3 path:
s3://project-bucket/
I have installed Big Data Tools and connected my aws account using the AWS plugin. Im wondering if its possible to move the entire project directory or point the part of my project directory with the large files to the bucket for later processing. I have seen tools such as Boto3 in other posts but am wondering if there is an alternative method or easier way to do this.

Download zip file on google drive

I want to create a MERN stack application from where I can auth to google drive, upload files and download zip folders. But I want to upload files in unzip format and when I want to download folder from google drive I want similar behaviors as google drive does i.e: It first zip the folder and the download it.
I checked it's documentation various npm modules but I am not able found and figure out how to do it.
Any one can help me in this

I want to configure AWS S3 on my django project but many of my client's media files are on my pythonanywhere server already

I want to integrate AWS S3 in my django project and i have successfully integrated that but the problem is many of my client's media files are already on the pythonanywhere server and now i want to change it to AWS S3 so the upcoming media files should be stored in S3. But now what happens is that the new files are storing correctly and i can open them but the previous files, when i open them because of the media files url i suppose django looks it in AWS S3 but they are not there. So i need django to understand to look for the media files where they are stored like for the previous files i want django to look in the server directory and for new files in S3 instance.
Hope that makes sense.
This is my AWS S3 configurations file for my django project
This is my django settings file, mainly focuses on media files
So if you want to move your static files that were uploaded before you configured AWS S3 with django you can run this command in your command line: python manage.py collectstatic.

Download Compressed Folder from Jupyter Notebook in GCloud Deep Learning VM

This seems to be a very simple question, but I couldn't find a way to do it. The jyputer notebook has the option to download file one by one. But my training process generates too many files, and I want to download them all at once. Is there any way to do it?
Assuming it is JupyterLab what you are using:
Open a new Launcher (+ icon) and start a new terminal session.
Use zip -r FILE_NAME.zip PATH/TO/OUTPUT/FOLDER/ to compress the required folder.
Download the zip file as you were doing with the other ones.

how to store image, video, audio file in aws server permanent outside tomcat deployed directory?

Currently, I am using AWS server as a beginner.
currently i am storing image, video file in project directory.
tomcat8/webapps/ROOT/Video/
but whenever project is redeployed all uploaded file is remove because of project WAR file is replace.
but i don't want to store these inside of AWS root directory. I want to store these into outside project directory.