How to execute a command on an S3 file [closed] - amazon-s3

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I have some video files stored on S3, and I want to get video information on them using ffmpeg. However, when I do a command such as :
$ ffmpeg -i 'http://test.s2.amazonaws.com/video.mov
I get a HTTP error 403 Forbidden response. How would I do this command? I also want to make sure not anyone can execute stuff on these files. Thank you.
Update: I was able to do this after making the ACL public-read=everyone for the video, didn't need to use s3fs after all.

You have two options:
You can download the file and run ffmpeg on localfile,
or you can use s3fs to mount your s3 bucket as a filesystem.

Related

Chmod permission of specific file extension under a folder via SSH [closed]

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
My folder is /images/. There are .png, .gif and .jpg images in this folder. How do I change permissions of specific file extension .jpg to 644 under the folder via SSH?
Connect to the remote computer with ssh and then execute this command may work:
find /images/. -name "*.jpg" | xargs chmod 644

Amazon S3 file upload via torrent [closed]

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I know that amazon-S3 supports BitTorrent and that we can download files using a torrent client. My question is,
Is it possible to upload files from my pc to s3 via torrent, either directly or using ec2 ?
Note:
I have a website where users upload large size video files which are stored in S3. It would be helpful if they could just upload a torrent file so that they can seed whenever they want and also have multiple seeders for the same file which decreases their upload time...
You'll have to install a BitTorrent client on EC2 (or on your own system), download the torrent file, and upload it to S3. S3 does not natively support fetching BitTorrent files from other sources and storing them in a bucket.
you can set a PUBLIC s3 file as downloadable by torrent if you add ?torrent at the end of the url.
For example, I just now uploaded following file:
http://mybucket.s3.amazonaws.com/Install/hexagon4.png
if you set the link to
http://mybucket.s3.amazonaws.com/Install/hexagon4.png?torrent
it will be downloaded as torrent.
I don't know if this is possible to do for closed files

How to transload a file from URL to Amazon S3 [closed]

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I have a 7GB file I need to download and store in a S3 bucket.
Can I transload it directly to S3 without having to download it to my computer?
AFAIK this is not possible. Here's another to confirm this: Is it possible to upload to S3 by just providing a URL?
You can make it appear that your files are using a different server to be accessed from or uploaded by playing with the CNAME record though: Using amazon s3, upload files using their servers but URL should appear to be from mine
But I don't think this is what you want.
You could however download the file to one of Amazon's EC2 servers and upload from there to S3.

scp error when copying files [closed]

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 11 years ago.
Improve this question
I am trying to use the scp command but I'm getting the following error:
permission denied
lost connection
I get this when I use scp to copy a file from a master node to a slave node.
This is probably more like your error, ya?
Permission denied (publickey).
lost connection
The simplest problem is that you don't have permissions to access the directories specified on at least one end of the copy operation, or you may just have a bad login.
You could try using sudo chown -R
*username* to make sure you have rights to the target directories on
both ends.
Check your directory specifications, make sure that you are using ~/Documents for instance instead of /Documents. Very different locations.
If you've got a bad login, I can't help you there, sorry. Try just ssh-ing into the target(s) make sure you can.
You could check out the manpage for more help
I think error in giving permission to folder which has the authorized_keys file
use chmod and set the permissions

Amazon S3 viewing files [closed]

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
Hi I have recently set up an amazon s3 account for a personal project.
I have successfully uploaded some image(jpg) files and have set the ACL for the bucket to public however when trying to view a file via the browser, the following xml is returned instead of the jpg.
<Error>
<Code>AccessDenied</Code>
<Message>Access Denied</Message>
<RequestId>77293A7937279B15</RequestId>
−
<HostId>
cQ3FXKg7ZU4z80QqUGMBheG0FRrFJP4HQx1pCy6UTFDk4pbjR8oYuCa1BmS6jnpe
</HostId>
</Error>
Am I missing something here? Do I need to set up a distribution, or should I be able to access the files regardless?
Any hints much appreciated.
You need to set the ACP of the resource (object) to anonymous ("public"), as well.
Note that this is not a programming question. If you need S3 support, check out the Amazon S3 forum.