Where Azure Log Analytics store the data - azure-storage

I started to study Azure Log Analytics and I'm wondering a very simple question: where are stored the data?
Is there a kind of database behind this resource? How can I access that?
If not, is there a way to "redirect" the logs into a particular storage?
I didn't find any info on the documentations.
Thanks

Azure Diagnostics is an Azure extension that enables you to collect diagnostic data from a worker role, web role, or virtual machine running in Azure. The data is stored in an Azure storage account (you have to assign a diagnostic storage account to store log data) and can then be collected by Log Analytics.

Related

migrating logins to another subscription in azure

Is there a way to migrate logins, including their sid and passwords, to a master on a server on another subscription in azure? There seems to be a couple ways to do this on-prem to on-prem but haven't found a way to migrate logins across subscriptions from azure sql to azure sql
Transferring an Azure subscription to a different Azure AD directory is a complex process that must be carefully planned and executed. Many Azure services require security principals (identities) to operate normally or even manage other Azure resources.
Steps to prepare for the transfer are in this link https://learn.microsoft.com/en-us/azure/role-based-access-control/transfer-subscription

What is the best method to sync medical images between my client PCs and my Azure Blob storage through a cloud-based web application?

What is the best method to sync medical images between my client PCs and my Azure Blob storage through a cloud-based web application? I tried to use MS Azure Blob SDK v18, but it is not that fast. I'm looking for something like dropbox, fast, resumable and efficient parallel uploading.
Solution 1:
AzCopy is a command-line tool for copying data to or from Azure Blob storage, Azure Files, and Azure Table storage, by using simple commands. The commands are designed for optimal performance. Using AzCopy, you can either copy data between a file system and a storage account, or between storage accounts. AzCopy may be used to copy data from local (on-premises) data to a storage account.
And also You can create a scheduled task or cron job that runs an AzCopy command script. The script identifies and uploads new on-premises data to cloud storage at a specific time interval.
Fore more details refer this document
Solution 2:
Azure Data Factory is a fully managed, cloud-based, data-integration ETL service that automates the movement and transformation of data.
By using Azure Data Factory, you can create data-driven workflows to move data between on-premises and cloud data stores. And you can process and transform data with Data Flows. ADF also supports external compute engines for hand-coded transformations by using compute services such as Azure HDInsight, Azure Databricks, and the SQL Server Integration Services (SSIS) integration runtime.
Create an Azure Data Factory pipeline to transfer files between an on-premises machine and Azure Blob Storage.
For more details refer this thread

Connecting storage securely to Azure Data Lake Analytics or Data Factory

I am setting up a new Azure Data Lake Analytics (ADLA) PAAS service to run USQL against some existing data sets in blob storage. The blob storage is firewalled for security and when I try to add the storage account to the data sources in ADLA I get the following error. Similar happens for data factory.
InvalidArgument: The Storage account '' or its accessKey
is invalid.
If I disable the firewall, the storage account can be successfully added. I have tried to add the relevant Azure Data Center IP Address ranges but the connection still fails. I have also ticked the "Allow trusted Microsoft Services box" but this does not seem include data lake or data factory. How do I access my storage account from ADLA but still have it secured?
You could install a selfhosted IR to access your blob storage. Whitelist the IP of the machine hosting your selfhosted IR.

Is There a Local Emulator for the Azure Data Lake Store

When developing for Azure storage accounts, I can run the Microsoft Storage Emulator to locally keep Blobs, Queues, and Tables without having to connect to Azure online.
Is there something equivalent for the Azure Data Lake Store? It would be nice to develop locally for a while without having to connect to Azure online.
Have you tried Visual Studio with the Azure Data Lake Tools plug-in?
As pointed out by David, you can develop Azure Data Lake Analytics (ADLA) projects locally without needing connectivity to Azure for the ADLA account or the associated Azure Data Lake Store (ADLS) account. Is there some other application you would like to use with ADLS?
Thanks,
Sachin Sheth
Azure Data Lake team
Same problem here.
AFAIK the Storage Emulator is not yet able to really handle Data Lake (ADSL Gen2) Requests.
This Uri works (but looks for a file, not a dir):
http://127.0.0.1:10000/devstoreaccount1/packages-container/Dir/SubDir?sv=2020-04-08&se=2022-10-13T14%3A43%3A39Z&sr=b&sp=rcwl&sig=d2SxwYCkJGyx%2BHac9vntYQZOTt5QVs1bKgKb4%2FgcQ9k%3D
This one doesn't:
Error: Status: 403 (Server failed to authenticate the request. Make sure the value of the Authorization header is formed correctly including the signature.)
ErrorCode: AuthorizationFailure
http://127.0.0.1:10000/devstoreaccount1/packages-container/Dir/SubDir?sv=2020-04-08&se=2022-10-13T14%3A43%3A39Z&sr=d&sp=rcwl&sdd=2&sig=KU%2Fcu6W0Nsv8CucMgusubo8RbXWabFO8nDMkFxU1tTw%3D
The difference is that the second one uses the resource 'sr=d' (directory) while the first uses 'sr=b' (blob).
Both items are working on real Azure Storage (with ADSL Gen2).
The request is already tracked here: https://github.com/Azure/Azurite/issues/553
Tested on VS 2022 17.3.6 using Server: Azurite-Blob/3.18.0

How to get error log in windows Azure server?

I have developed test application to display claims of authenticated identity.
this application is working in local but when I publish to windows azure it gives some dot net error. Can Anybody explain me how to get error log in windows azure server?
Thanks in Advance !!!
I suspect you aren't looking for help with Azure Storage - or just in case you are I have included a detailed explanation below. For Azure log information see here: Windows Azure PaaS Compute Diagnostics Data.
If it is storage you are interested in, the following blog post provides a good overview of the logging capability: Windows Azure Storage Logging:Windows Azure Storage Logging: Using Logs to Track Storage Requests.
Jason