Is There a Local Emulator for the Azure Data Lake Store - azure-data-lake

When developing for Azure storage accounts, I can run the Microsoft Storage Emulator to locally keep Blobs, Queues, and Tables without having to connect to Azure online.
Is there something equivalent for the Azure Data Lake Store? It would be nice to develop locally for a while without having to connect to Azure online.

Have you tried Visual Studio with the Azure Data Lake Tools plug-in?

As pointed out by David, you can develop Azure Data Lake Analytics (ADLA) projects locally without needing connectivity to Azure for the ADLA account or the associated Azure Data Lake Store (ADLS) account. Is there some other application you would like to use with ADLS?
Thanks,
Sachin Sheth
Azure Data Lake team

Same problem here.
AFAIK the Storage Emulator is not yet able to really handle Data Lake (ADSL Gen2) Requests.
This Uri works (but looks for a file, not a dir):
http://127.0.0.1:10000/devstoreaccount1/packages-container/Dir/SubDir?sv=2020-04-08&se=2022-10-13T14%3A43%3A39Z&sr=b&sp=rcwl&sig=d2SxwYCkJGyx%2BHac9vntYQZOTt5QVs1bKgKb4%2FgcQ9k%3D
This one doesn't:
Error: Status: 403 (Server failed to authenticate the request. Make sure the value of the Authorization header is formed correctly including the signature.)
ErrorCode: AuthorizationFailure
http://127.0.0.1:10000/devstoreaccount1/packages-container/Dir/SubDir?sv=2020-04-08&se=2022-10-13T14%3A43%3A39Z&sr=d&sp=rcwl&sdd=2&sig=KU%2Fcu6W0Nsv8CucMgusubo8RbXWabFO8nDMkFxU1tTw%3D
The difference is that the second one uses the resource 'sr=d' (directory) while the first uses 'sr=b' (blob).
Both items are working on real Azure Storage (with ADSL Gen2).
The request is already tracked here: https://github.com/Azure/Azurite/issues/553
Tested on VS 2022 17.3.6 using Server: Azurite-Blob/3.18.0

Related

How to Connect ADLS Gen-1 with Azure ML Studio

Want to connect ADLS Gen-1 with AzureML Studio.
I try to find out some solution but could not get
Direct method:
Currently, Azure Data Lake Store is not a supported source.
I would suggest you to vote up an idea submitted by another Azure customer.
https://feedback.azure.com/forums/327234-data-lake/suggestions/15008490-adl-store-connector-for-ml-studio
All of the feedback you share in these forums will be monitored and reviewed by the Microsoft engineering teams responsible for building Azure.
By using the Import Data module, you can access data from one of several online data sources while your experiment is running:
• A Web URL using HTTP
• Hadoop using HiveQL
• Azure blob storage
• Azure table
• Azure SQL database or SQL Server on Azure VM
• On-premises SQL Server database
• A data feed provider, OData currently
• Azure Cosmos DB
For more details, refer “Supported data types in Azure ML studio”.
In-direct method:
Azure Data Lake Analytics can also be used to write data out to Azure Blob Store, and so you can use that as an approach to process the data in U-SQL and then stage it for Azure Machine Learning to process it from Blob store. When Azure ML supports Data Lake store, then you can switch that over.
For more details, refer "How to use ADLS as an input data set for Azure ML Studio".
Hope this helps.

Azure Gov Cloud and Azure Functions trigger on Storage

I have hard time with Azure Functions on Azure Government. I need to create a C# trigger bases process on Azure Storage. The goal is to automate the process of the loading the files into Azure SQL DB when a file is dropped into Azure Storage.
Since Azure Functions in Azure Government are not fully comparable to Azure Function in regular Azure and not all UIs are the same, I can't deploy the function to trigger on a storage file.
I was able to build the process in regular Azure Cloud following instructions from https://github.com/yorek/AzureFunctionUploadToSQL but since Azure Government is missing the UI for Azure Functions I'm having hard time to replicating the process in Azure Government.
Portal UI support is not yet available in Azure Government, but it is coming soon. Additionally, Azure Government currently supports "App Service plan" ("Consumption plan" coming soon).
In the meantime, you can do everything you need. First, provision your Azure Function in Azure Gov via the Azure CLI by following this Quickstart example for Functions on Azure Gov. That same link also shows you how you can use Visual Studio to set up your triggers (in your case, a Blob trigger).
Once complete, deploy your Function to Azure Gov with Visual Studio.

Sand box environment for Data Lake Store and Analytics

Is there any sand box environment for Data Lake Store and Analytics so that I don't have to use my Azure Credits?
Azure Data Lake Analytics (ADLA) does have a mode for local execution. You install an emulator and this enables you to run your U-SQL scripts from Visual Studio either against your local instance or your Azure ADLA account.
Some reading on the topic:
https://azure.microsoft.com/en-gb/blog/run-u-sql-scripts-locally-with-updated-azure-data-lake-tools-for-visual-studio/
https://learn.microsoft.com/en-us/azure/data-lake-analytics/data-lake-analytics-u-sql-sdk
You can use the ADL tool with VisualStudio Community Edition which gives you the experience for free.

How to write sqlcmd results directly to Azure Storage using Azure PowerShell?

Current story:
Moving overall BI solution fully to Azure cloud services. Building a new Azure DW and loading data from an Azure DB. Currently, Azure DW doesn't support linked servers and/or the elastic query (this is only supported in Azure DB). Due to price, we can not use data factory or an instance of SSIS. We can't use bcp as we don't have a local directory to hold the file in between loads.
Is it possible to use Azure PowerShell with sqlcmd to write results of a query directly to Azure Storage, without having to write to a file on a local directory in between?
Are there other options that aren't mentioned above?
Thank you for any input.
The current Azure Storage PowerShell (Set-AzureStorageBlobContent) only support upload blob from local file.
Azure Storage Client Library (https://github.com/Azure/azure-storage-net) support to upload blob from stream, can you try to develop your own application with the Azure Storage Client Library?
If your data is big, you can also try https://github.com/Azure/azure-storage-net-data-movement/, it has better performance in upload big blob.

How to get error log in windows Azure server?

I have developed test application to display claims of authenticated identity.
this application is working in local but when I publish to windows azure it gives some dot net error. Can Anybody explain me how to get error log in windows azure server?
Thanks in Advance !!!
I suspect you aren't looking for help with Azure Storage - or just in case you are I have included a detailed explanation below. For Azure log information see here: Windows Azure PaaS Compute Diagnostics Data.
If it is storage you are interested in, the following blog post provides a good overview of the logging capability: Windows Azure Storage Logging:Windows Azure Storage Logging: Using Logs to Track Storage Requests.
Jason