Dashboard Backups in Datadog - backup

Is there a Backup and Restore feature coming for Datadog Dashboards
(Not delete recent feature)
Roadmap answer from PM

Related

ejabberd Mnesia database backup

Does making an ejabberd binary backup of mnesia database from the admin panel will also back up archived messages stored in MUC archive and private chats archive? If not, how to back up archived messages?
ejabberd Mnesia backup backups all the data stored in Mnesia, so if your archive are in Mnesia they will be backuped as well.
However, like always with backup, you must test the process from backup to restore to validate that it works as expected and matches your needs.

Backing Up SQL Server To Cloud?

To start i dont have much experience with DB's. In summary, i have an application for a client. They dont want to necessarilary host their DB online but have a local sql server set up for their 2 computers.
I have a batch script that backs up the database every night. Is there a way in the batch script to send them to the cloud like a skydrive,etc?
Try our SQLBackupAndFTP software. You can schedule backup jobs with SQLBackupAndFTP (full, differential and transaction log backups), save backups at local folders, FTP, Dropbox, Box, Google Drive, Amazon S3, SkyDrive, delete old backups and configure email notifications... Basic features are available in free version or you can try all features in trial mode.
Just backup to SkyDrive folder. If DB is big use rsync for sending to cloud or do full backups every week, differential every day and transaction every hour (depending of you application etc) then there will be less data to send to cloud.
backup database to SD folder - using backup compression (if available in your edition of sql server, see WITH COMPRESSION clause of BACKUP statement help in BOL) or using 3rd party backup compression tool (either free or paid)
also you can backup database to temporary folder and then zip it to SD folder
If you are willing to spend a bit of money, cloudberry has an sql backup tool that will do exactly that (and more).
http://www.cloudberrylab.com/sql-backup-amazon-s3-azure.aspx

Migrating SQL Server database from AWS to Azure

I have a large database in an AWS instance running SQL Server 2008 on a Windows Server 2008 R2.
The database is constantly changing and writing information, and its size is about ~100GB
I wish to migrate from our Amazon services to Microsoft Azure.
But I cannot afford any lost of information more them for more than 20-30 minutes
I don't mind using the Azure SQL or running a SQL Server under a VM in Azure Cloud, but I must keep the databases live and updated, there are few main tables that information is being added to them constantly
What would be the best way to do so ?
if you are using an AWS instance and not RDS and you are going to an Azure instance and not "Azure SQL Database" you can use log-shipping or something similar to get the downtime down to a few seconds: http://msdn.microsoft.com/en-us/library/ms187103.aspx
The steps you need to take:
Take full backup on AWS
restore full backup without recovery on Azure
take log backup on AWS
restore log backup without recovery on Azure
repeat 3 and 4 until the time it takes is short enough (you probably want to script this out)
take app offline
take another log backup on AWS
restore that log backup WITH recovery on Azure
repoint App to Azure
bring App online again.
3, 4 and 5 is what log-shipping would automate, but you could just write a powershell script too.

Team Foundation Server 2012 Update 2 - Scheduled Backups Wizard

After applying Update 2 - the Scheduled Backup is now a build-in feature (not requiring Power Tool). I have difficult understanding the required permissions required to dump a backup on a file-share. Even when trying to dump the backup on a local share - \mytfs2012server\c$\Backup - the wizard ends up with a:
TF401002: The SQL Server Database Engine failed to save the database backup to path... Please grant SQL service account read/write access to that folder.
The message is very simple. I have tried different things - without success. With Update 1 + Power Tools I succeeded to use the wizard after some tweaking. Anyone that can list the permissions required for the Scheduled Backup for Update 2?
Edit:
Update 3 did not change anything for me. However it looks like Update 4 will. From this page it looks like Update 4 will include changes regarding backup:
Scheduled Backups configuration is no longer blocked if the SQL Server service for TFS 2012 is running as a virtual account (for example: NT Service\MSSQLSERVER).
In TFS 2012 Update 2 and Update 3, transactional backups record a failure when they try to run while a full or differential backup is running.
Scheduled Backups no longer run transactional backups if a full or differential backup is running. Instead, the job will be suspended until the other backup has finished running.
I had the same Problem.
Infrastuctur: TFS 2012 Update 2, SQL Server 2012
Solution:
I created a simple shared Folder like \BACKUPSERVER\TFS_Backup$ on the DB Backup Server.
I configured the TFS admin and the SQL Admin with Full Control in the Security and Share Permissions.
Now the Backup Job is running.
But now i have an other Problem:
After running a Full Backup the Transaction-Logs were not truncated, does any one have the same Problem or maybe an solution?
Are you using virtual accounts for the server? Something like 'NT Service\MSSQLSERVER'? If so be aware that they aren't supported in the Scheduled Backups Wizard in both TFS 2012.2 & 2012.3.
You either have to create the backup plans manually or switch the SQL service account to a domain account or Network Service.
EDIT:
As per the comments, TFS 2012.4 RC2 and up has removed this restriction

Azure Bacpac file Restore Error

I do the daily bacpac file backup from Azure database using RedGate Azure Backup tool.
Until 31st of May, All my Bacpac files can restore to Local SQL2008R2 database using DAC Client Tool.
Starting from 1st of June, I got error like:
Failure Creating schema objects in database
'XXXXX' System.Data.SqlClient.SqlException (0x80131904): Cannot
find the object "dbo.XXXX" because it does not exist or you do not have per
missions.
When I check it table still exist in DB.Also No idea for this.
Is there anyone face that kind of missing table, invalid Column.. etc
RedGate Azure Backup is deprecated now. The DAC Client Tool (DacCli.exe) has been integrated into SSMS 2012 and SSDT. You could try updating your DAC Client Tool to the latest version (1.6 from 1.2)
SSMS 2012 Data Tools now allow you to Import Data-tier Applications. Take a look at the Data-Tier Application Framework Feature in SQL 2012 downloads. View this MSDN article for reference.
I use the free utility SQL Backup and FTP to backup Azure SQL Databases to a local .bacpac file. The free version allows you to backup up to 2 different databases and then the paid versions are very reasonable (i.e. Professional version is $70 or so).
http://sqlbackupandftp.com/
It actually runs on a schedule for me, automatically emails me the results of the backup and stores it on my Google Drive (even with the free version - I just have it store the timestamped backups in the file folder of my Google drive).
The procedure I use to restore it to a local database is:
Back up SQL Azure to .bacpac file
Login to local database server as sa
Right-click Databases
Click Import Data-Tier Application
Select .bacpac file either from Azure or local disk
Enter the name and the locations to store the data and log files
If necessary, set the original database to single user mode and delete it
If necessary, rename restored database to original database name
Note: Must use SQL Server 2012 (or later) management tools