Where did expiring S3 URLs go in the AWS console? - amazon-s3

I used to be able to create expiring (signed) URLs directly from within AWS console very easily, just right-click the item, choose web URL, and voila.
This feature seems to have disappeared sometime this year. Does anyone know how to do it now? I am not looking for a programmatic solution, I just occasionally need to whip up an expiring URL for a file, about once or twice a year. I really don't want to bother with a third party tool either, I want to know how one does it from within AWS console now, in 2015.

This option was never available in the AWS Management Console.
If you use Actions -> Open on the file from within the Console it generates a signed URL, but there is no capability to edit the parameters.
It is, however, available in:
The AWS Toolkit for Visual Studio
Cloudberry Explorer PowerShell
Bucket Explorer
GitHub amazon-s3-url-signer
GitHub s3-signed-url
Your own code! Signing Amazon S3 URLs

Related

serve oracle service cloud Customer portal locally?

I am working on customizing the oracle service cloud customer portal, but since OSvC provides only WebDAV to connect to it. It is very time-consuming to edit files and then upload them to WebDAV even for a single word change.
I am looking for a solution to serve it locally make desired changes and then upload the desired code to webDEV.
But after searching the file structure I can not make which framework it uses, I tried to use websites like https://builtwith.com/ and WhatRuns but they are also not able to find anything useful.
Although after searching in the file structure, I find some files of CodeIgnitor but the structure is way more different than the CodeIgnitor folder structure.
The short answer is no, you will not be able to run Customer Portal locally. While it is a fork of CodeIgniter from many years ago, there are server-side dependencies that will prevent you from running it in a local sandbox.
That said, it is possible to automate many of the manual tasks of interacting with WebDAV for change testing. If you edit locally, then you can use scripting hooks or event RPA robots to automate some of the manual file movement. Personally I have a flow to edit remotely in my test environment with an editor (like VSCode or Nova) that can connect to a remote server via WebDAV and edit files directly in the development area of a site. Then, when finished, I have a script that pulls down the latest version of all files and then allows me to commit changes to Git for SCM.
Another option is RPA. You can develop a robot that can be run to automate the manual tasks that you face in your workflow. Personally, I think that scripting is a better solution than RPA since you can automate all of the actions via scripting or a shell. But, it's another option to consider.
Another way of "Live editing" the OSvC CP code is to connect to WebDav via a software that supports it like Mountain Duck which uploads your code to OSvC on save.
OR use the better solution Windows Explorer which supports connecting to WebDav and treating it like a network drive, by going on My Computer -> Computer -> Map Network Drive then put https://yoursite.custhelp.com/dav/cp click Next then you'll be promoted to login using your OSvC login.

AWS download files from S3 in web browser

I am a newbie to AWS and one of the tasks I have is to figure out how to download MSIs, ISOs stored in S3 through a web browser. I read that I could use CLI behind the scenes. So if a customer clicks on one download; the app would make a request to S3 using one of the commands and that would download the file lets say through Google Chrome or IE (Please correct me if I'm wrong in the usage of CLI).
Now if the download stops for some reason due to internet failure; is there a way to resume the download? How do I get a download done through a client.
Thanks in advance for helping. Unfortunately the AWS links gave me very little information so seeking help here!
May
Files stored in Amazon S3 can be directly accessed via web browser, just like clicking a link on any website.
If the files are marked as publicly-accessible, anyone with the link can download the file.
If you wish to limit access to the files, your application can generate a pre-signed URL that will work for a limited time period that you specify (eg 5 minutes). Users can use/click that link to download the file within that time period.
You can also download files using the AWS Command-Line Interface (CLI), which has Copy and Sync commands. This would, however, require installation of the CLI on the user's computer. This is great if they are regularly download files or if you wish to automate the download (eg every hour or daily).
If you wish to explore AWS, sign-up for an account and make use of the Free Usage Tier, which lets you try some services for no charge.

How to browse source code in developer console on Google Compute Engine

I have used Click to deploy MEAN Stack on Google Compute Engine. Everything is fine, runs in the Cloud and on my local machine.
When I try to browse source code in the developer console, I get the Getting startet screen for cloud source tools.
Should it not be set up automatically when I created the instance?
When I SSH into the instance I can see the content of opt/myApp.
This I would like to see in browse source code also.
If I try to git clone the Cloud Repository it is empty.
I'm currently using the trial version. Maybe this is the reason I can not browse or clone the Cloud Repository?
This is a really good suggestion. It would be very interesting to integrate click to deploy in more Google Cloud Platform services. The MEAN click to deploy doesn't have much of a sample application installed, which is why it doesn't make sense to put it in a code repository. The application stack itself (mongo, express, angular, node) you wouldn't put in a repository, which is really the only thing installed on the MEAN Click to Deploy server.

TCL Amazon S3 Interaction

I have a TCL/TK Windows application that creates a small executable that I distribute to my customers. Because it is an exe file I can not email the file. Instead I upload it to an Amazon S3 bucket then create a URL link and email the link to them. They download the file from the link and run the exe.
What I would like to do is add the ability upload to an Amazon bucket within the application that will enable me to upload the file and create a URL that I can copy and email to the customer. I have seen Amazon S3 API's written for other languages, python, java, but not TCL. Has anyone done this? How hard is it? Can you point me to a tutorial?
Actually I do not have to use a S3 bucket. If there is another suggestion for how to distribute small files to customers from within TCL programs I am open to suggestions. Besides what has been laid out above the only other requirement is that multiple people must be able to upload to the same location, the TCL program runs on Windows and I would like to not use a 3rd party program. Security is not a major concern, nor is privacy, these things are handled other ways.
Actually, Tcl does provide an S3 package, but since I don't have Amazon S3 account, I cannot test it out.

Updating permissions on Amazon S3 files that were uploaded via JungleDisk

I am starting to use Jungle Disk to upload files to an Amazon S3 bucket which corresponds to a Cloudfront distribution. i.e. I can access it via an http:// URL and I am using Amazon as a CDN.
The problem I am facing is that Jungle Disk doesn't set 'read' permissions on the files so when I go to the corresponding URL in a browser I get an Amazon 'AccessDenied' error. If I use a tool like BucketExplorer to set the ACL then that URL now returns a 200.
I really really like the simplicity of dragging files to a network drive. JungleDisk is the best program I've found to do this reliably without tripping over itself and getting confused. However it doesn't seem to have an option to make the files read-able.
I really don't want to have to go to a different tool (especially if i have to buy it) to just change the permissions - and this seems really slow anyway because they generally seem to traverse the whole directory structure.
JungleDisk provides some kind of 'web access' - but this is a paid feature and I'm not sure if it will work or not.
S3 doesn't appear to propagate permissions down which is a real pain.
I'm considering writing a manual tool to traverse my tree and set everything to 'read' but I'd rather not do this if this is a problem someone else has already solved.
Disclaimer: I am the developer of this tool, but I think it may answer your question.
If you are on Windows you can use CloudBerry Explorer Amazon S3 client. It supports most of the Amazon S3 and CloudFront features and It is freeware.
I use the Transmit Mac app to modify permissions on files I've already uploaded with JungleDisk. If you're looking for a more cross-platform solution, the S3Fox browser plugin for Firefox claims to be able to modify permissions on S3 files as well.
If you need a web based tool, you can use S3fm, free online Amazon S3 file manager.
It's a pure Ajax app that runs in your browser and doesn't require sharing your credentials with a 3rd party web site.
If you need a reliable cross-platform tool to handle permissions, you can have a look at CrossFTP Pro. It supports most of the Amazon S3 and CloudFront features as well.