Automated incremental backups from Plesk on Centos to Amazon S3 - backup

I've done a far bit of research on this via Google and there seems to be quite a few ways of possibly doing this.
I'm looking to incrementally backup new and updated files in two directories on my Plesk run Centos 5.2 server: /backups and /var/www/vhosts (preferable only httdocs within each vhost)
Has anyone got some great feedback from using the various solutions - seems to be various Java, Perl and Ruby based solutions out there.
Many thanks,
Chris

I use http://s3tools.org/s3tools and the tool S3cmd in a cron job to upload all the stuff I want to back up. I know of people that use SVN and then just back that up.

Yes - the best solution I've found is Jungle Disk
https://www.jungledisk.com/
it's a snap to set up!

Related

Dropbox sync bandwidth is it limited?

I have installed dropbox python client for linux and I noticed the sync bandwidth is quite limited:
$ dropbox status
Syncing (252,088 files remaining, 18 days left)
Downloading 252,088 files (35.1 KB/sec, 18 days left)
Is there a way to make it faster?
Note: Yes I have a 100Mbit/s internet connexion...
Firstly, check if there is a 75% cap enabled, as mentioned here
If there isn't then it's probably your Internet, try switching to a different network source (from wireless to wired) or use a different Internet connection. I had the same issue before and it was solved by changing to a different Internet connection, yes I have 100Mbit/s too but it didn't help.
Alternatively
If you already have another synced up dropbox, just copy the files over to the new install of Dropbox, if you're just trying to get the initial sync done.
Also take a look at LAN Sync, a feature in Dropbox
This honestly isn't a SO question because is isn't really a programming question, a forum like Superuser.com might be better suited perhaps.
edit: saw that you already have a superuser account, my bad. :)

handling file upload and serving in a distributed web application

I'm going to deploy a web application with multiple Pyramid application servers and nginx as a load balancer.
This application will have a feature for uploading files which should be available for downloading afterwards.
Total size of uploaded files may be very big so I'd like to deploy a separate file webserver to serve these static files. (this is one reason why I don't like rsync solution proposed here).
What is the best solution to handle file upload and syncronization in this case? I was thinking about NFS or something like that, but I'm not sure it is a good way to solve the problem. I suppose there must be some best-practices here or even a tool or library for these purposes.
UPDATE:
I don't want use cloud services like Dropbox, it would be nicer to find some syncronization solution inside the network segment.
UPDATE2:
I finished with setting up NFS, for now it works perfectly.
not really a python or pyramid related question. But, you should investigate distributed file systems and CDN's both of which are for this kind of thing. gridfs is easy enough to get going with. But there are plenty of other options. Both Amazon and Google have similar services.

online backup solution with api for desktop

I made a small backup application that simply creates an archive out specified files and folders. Now I need an online service to backup that online. Which service can i use that can be integrated into my app ?
Options I considered:
dropbox is ideal, but they have all but abandoned the desktop.
skydrive has no api.
I couldn't find any free reliable backup service that uses ftp .
anything else ? it should provide 1-2 gb of free space and be reasonably reliable.
Thanks
My app is in C#, but can be ported to any other language as well..
In your case, Amaxon's S3 seems more fitting but that's not free.
Depending on your target audience, you can create a local archive and have that picked up by your regular backup solution. You might try Wuala,or SpiderOak. Expand Wuala by adding your own space. Spideroak is free up to 2GB (more if you invite friends), and also provides a good alternative to Dropbox (if you want to see how to migrate from dropbox to spideroak see my blogpost about that).
Try box.net, now known as box.com or simply Box
reference: http://developers.box.com/docs

Capistrano deployment with lots of images

So we have this basic Rails 3 website with capistrano 2.5.19 plus multi-stage extension.
The site is simple, but it has 40,000+ of images out there. So deployments take a long time, going both to our QA server and production. The issue is not usually network load, because capistrano only downloads what changed in svn. The issue is the time it takes for our servers to backup the old release (40k worth of images) and copy the new release (another 40k of images.)
Does anyone know of a best practice approach to this? Is the only way to split this into two SVN folders and two deployment scripts combined with some symlink magic? Or can i tell capistrano to exclude the images on certain deployments where I know images have not changed?
Well, we have this issue too. A solution is a library called fast_remote_cache if you're on linux.
https://github.com/37signals/fast_remote_cache
The idea is that it hard links to the cache so the copy is much faster. Once the site finally gets large enough that even this takes too long, then it is time to consider asset servers.
It's probably better not to have all those images in your repository, or at least in a different repository.
You'll want to see about setting up an asset server. They're easy to hook into Rails, as long as you use the XXX_tag helpers. And you could just have the asset server run plain old Apache - not need for anything dynamic on it...
You might also be able to hook a "cloud" file store (I'm thinking Amazon S3, but there are plenty of others) in to serve the same purpose - they'll provide file backup (and version control, in some cases), and you won't even have to worry about running the asset server yourself.
Hope this helps!

Apache 2 Administration Tools on Ubuntu or Windows

I am totally new to Apache and Ubuntu, is there an admin tools where the available functions and features and settings are more accessible? I am finding it very difficult to do even the simplest things.... please help.
Cheers.
apacheConf is available for most flavours of Linux, however:
1) it ignores the distribution-specific layout of the config files and implements its own structure
2) its just a tool for updating the config - it doesn't actually tell you what the config options mean
3) like any gui tool it abstracts the actual data (in this case, the apache config) you are manipulating
4) there are at least 3 different programs branded as apacheConf
http://www.apache-gui.com/apache-linux.html
http://www.redhat.com/docs/manuals/linux/RHL-7.2-Manual/custom-guide/apacheconf.html
and the config tool bundled in Mandriva
5) it limits what you can actually configure to the design on the UI (e.g. RedHat apacheConf only allows you to set one listening address
Having buttons to click does not help you understand what you are actually trying to achieve (unless your objective is an MCSE of course ;). I'd recommend you go get a good book on Apache config and take the time to read it.
C.
Try Webmin - a nice web-based GUI tool. There is a .deb package for ubuntu in their downloads page.
http://webmin.com
Here are some options to consider for helping you to better understand how Apache works and how to administer/configure the server:
Read the official HOW-TO
Take a look at a commercial tool like ApacheConf
Ask questions here about the specific issues you are having
HTH,
-aj