Restore database from backup drupal 8 - backup

I accidentally deleted my drupal 8 files. But my database is not damaged right now. Is there anyway I can bring my site back to live.
What I did: I have my site old backup [only files]. Copied those backup to root folder replaced database credentials in settings.php. But no luck.
Any other way to bring back my site live from database.

If you have drush you try with
drush sql-cli < my_directory/my_drupal8_dump.sql
So, in you backup files just run this command, not need to update settings.php

Related

cPanel / WHM restore from backup tar archive (from another cPanel server)

I am quite new to server migrations, but fairly familiar with cPanel though. My current task is to migrate an entire website from a server with cPanel to another one.
What I did so far:
Use the Backup Wizard on the old server to create a full backup archive, and FTP it to the new server.
The full archive (about 6 GB because there are a lot of images) is now in my new server's public_html directory.
Now, what I need is a way to make the server take this tar archive, which is a full backup, and restore from it.
What I tried:
I tried simply extracting the archive, but it is taking forever to finish (again the archive is 6 GB), and for some reason my browser tab has to stay open until the end of the process, otherwise the extracting halts.
Also, as I have WHM access, I tried the "Restore a Full Backup/cpmove File" option, but for some reason, under the "Username for the account that you wish to restore:" textbox, WHM does not find my cPanel username.
If anyone can either tell me what I am doing wrong, or propose another option, I would really appreciate it.
P.S.: I only have WHM access to the new server, not the old one.
Edit: I got the WHM method working now. My mistake was that my tar archive was not stored in the /home directory, but in the home directory of the cPanel user (which is /home/username)!
Move your cpanel full backup file to /home directory and if you root have access of the server use below command as /scripts/restorepkg (cpanel username )
OR
Login to WHM with root user and go through the steps mentioned on below URL.
http://documentation.cpanel.net:8090/display/68Docs/Restore+a+Full+Backup+cpmove+File

Why does Dropbox break my Drupal site?

I'm using Apache 2.4 on Ubuntu 14.04 and Drupal 7. I wanted an easy remote backup for my Drupal site, so I tried putting it in Dropbox. Then, something happened, and Apache started giving 403 errors ("You don't have permission to access / on this server.") for the site.
I recovered from an old backup, but I still can't figure out what happened. I diffed the Dropbox directory and the backup, and they're the same. I reset all the permissions in the Dropbox directory to match the backup, but the version in Dropbox still won't work. I also tried copying the files from the Dropbox directory into the location of the backup, and that still didn't work.
I'm a bit at a loss as to what went wrong. Does anyone have ideas as to what Dropbox might have broken?
Rather than trying to run the site from a Dropbox location, why not just make Dropbox the repository for backups?
Try this:
https://www.drupal.org/project/backup_migrate_dropbox
This will also backup your database which you didn't mention as part of your Dropbox backup strategy.
Or you could run a cron task to run incremental backups via rsync to the Dropbox folder (Assuming you have already gone through getting Dropbox connected any syncing with Ubuntu). I use the process documented at this location and it has worked perfectly on a number of different web servers:
http://hadzimahmutovic.com/rsnapshot-mysql/rsnapshot-backing-mysql-databases

Uploading umbraco site on server from local host

I've successfully created site using Umbraco now its time to upload it on hosting server..
i've searched and got one paid product for the same..and i dont want to use it.
Has any body tried developing Umbraco site on local and then uploading it on server?
If yes then please help me doing that.
First I run the umbraco install from a local IIS website. Then I setup my visual studio solution for that website (and my souce control). Then I work, until I reach a beta version, then I go through this process for deploying:
Ftp over to the remove website and copy the whole website (I actually use Beyond compare).
Connect to my local database with management studio and create a .bak file.
Upload the .bak file to the database server.
Restore that database
Review connection strings in web.config
Then I'm pretty much done.
Once I'm "live" and have content I don't want to lose, when I want to work on the website, I bring back the live database through a .bak file, then I make my changes. They often include DB changes since the schema is basically in the database. I note all the operations I do. Once my changes are ready I manually replicate the changes on the live site as I update the files.
This is very painfull but I tried solutions like courrier and other things like that and they are not reliable enough for production I find. Manually is the only risk free way I see for the moment.
Hope this helps.
Yes, that happens all the time. Use FTP to copy your local installation to your webserver, modify the web.config to point to the correct database and your website should be up-and-running.
I'm sure there are more elegant solutions with less clicks but here's how I do it on azure websites with sql, not sure what hosting/db you're using:
1) Create an empty db on azure with the same login and user as my local db.
2) Create an empty site on azure connected to my db.
3) Download the publishing profile.
4) Upload the db the first time with Sql Azure Migration Wizard.
5) Import the publishing profile into and upload the site with WebMatrix.
6) Thereafter I deploy the site and db with WebMatrix.
WebMatrix uses WebDeploy or FTP, you can use WebDeploy through IIS if you like, and FTP.

cPanel - No permission to download backup

Thanks in advance for any help, you guys are awesome!
A previous web developer made a backup of the site in question to the server months ago, and the permissions for this backup are 600. I can't change the permissions or download the file, although I can still move it between folders.
I've read having low disk space is sometimes the problem, but in this case there are 2 GB left, so that should be fine.
There is an other way to download backup from server to local machine, if you are able to move backup file in folders then move it to your website's home directory like '/home/user/public_html/backup.zip' and then browse it in browser and download the backup file.
http://yourdomain.com/backup.zip
It will provide you download option.

EXECUTE master.dbo.xp_delete_file folder permission issue

I'm trying to run a Maintenance Cleanup Task to remove .bak files older than 2 days (simple enough).
Been trying all variety of .bak, BAK, .*., and editing the path, but the files are still not getting removed even though I receive a "job succeeded" log message.
I'm at the point where I believe it's a folder permission issue.
How do I make sure my SA has the proper permissions to remove files from a folder?
T-SQL:
EXECUTE master.dbo.xp_delete_file 0,N'D:\Program Files\Microsoft SQL Server\MSSQL10.MSSQLSERVER\MSSQL\Backup\',N'BAK',N'2012-06-21T11:35:59',1
Thanks.
I always have domain account with some (in your case sa) privilages on sql server. Then you add domain login to folder.