Azure Synapse Analytics (formerly SQL SW) vs Azure Synapse Analytics (workspaces preview) - sql

What are the differences between the following Azure Services?
Azure Synapse Analytics (formerly SQL DW)
Azure Synapse Analytics (private link hubs preview)
Azure Synapse Analytics (workspaces preview)
Are these three different products? Or are the two preview services just new features that will eventually be added into Azure Synapse Analytics?
The documentation is a little confusing. This FAQ (https://learn.microsoft.com/en-us/azure/synapse-analytics/overview-what-is) for the workspaces preview, for example, just looks like a FAQ for the overall Azure Synapse Analytics service.

It would be useful to link to a document mentioning these terms so I could have some context. Without context, this is my understanding of these:
Azure Synapse Analytics (formerly SQL DW)
This is just the MPP relational platform piece of "Azure Synapse Analytics"
You can connect to it using Azure Data Studio, SQL Server Management Studio, or Synapse Workspace and run SQL queries on it. It's a relational database that stores data across 60 shared-nothing nodes
Azure Synapse Analytics (private link hubs preview)
private link is a new feature across many Azure resources (data lake etc.) that allows you to confine connectivity to internal Azure VNets, meaning that you can use the resource without requiring public access. This feature is not specific to Synapse, it's a network connectivity feature being rolled across multiple azure components
Azure Synapse Analytics (workspaces preview)
This is the actual front end that has tabs for various analytics components. One component is the MPP platform that used to be called SQL DW. Another component is MS spark engine. Other components are Power BI and Data Factory.
Do you have a use case or an objective here?

Related

Can't Access Azure Synapse Spark Pool Databases on SSMS

Since I've starting using Azure Synapse Analytics, I created a Spark Pool Cluster, then on the Spark Pool cluster I created databases and tables using Pyspark on top of parquet files in Azure Data Lake Store Gen2.
I use to be able to access my spark Database/ parquet tables through SSMS using the Serverless SQL endpoint but now I can no longer see my spark Databases through the Severless SQL Endpoint in SSMS. My spark databases are still accessible through Azure Data Studio but not through SSMS. Nothing has been deployed or alter on my side. Can you help resolve the issue? I would like to be able to access my spark databases through SSMS.
Sql Serverless Endpoint
Azure Synapse Database
If your Spark DB is built on top of Parquet files, as you said, databases should sync to external tables in Serverless SQL pool just fine and you should be able to see synced SQL external tables in SSMS as well. Check this link for more info about metadata synchronization.
If everything mentioned above is checked, then I'd suggest you to navigate to Help + Support in Azure Portal and fill in a support ticket request with details of your problem so engineering team can take a look and see whether there is some issue with your workspace or not.

Unable to connect Azure purview from Azure Synapse

I followed the documentation for connecting azure purview account with synapse.
I do not see option - Manage -> Azure Purview in Synapse Workspace.
UPDATE: Azure Purview Integration is now available in Azure Synapse Studio.
On January 8th, 2021
The integration between Azure Synapse Analytics and Azure Purview is currently under Preview. If you are interested to try Azure Purview in Synapse, please connect with your Microsoft Sales Representative.
To get access to this feature you need to reach out to your sales representative at Microsoft.
In case, if you do not have a sales representative at Microsoft. I would suggest you file a support ticket and with a strong business use case.
Once you have created a support ticket, I would request you to share the support request number so that I can route it to the Azure Synapse team to enable access to the Azure Purview.
Hello as mentioned Pradeep, the integration feature is under gated preview but will be public very soon (end of January). In the meantime, if you still want to connect, please connect through my LinkedIn profile (Arnaud Comet).
Thank you,
Arnaud

How do I import one workspace into another in Synapse Analytics Workspace

How do I import one workspace into another workspace in Synapse Analytics? For example, I would like to import dev synapse into qa synapse.
It should behave like Azure Data Factory does where you can Import and Export ARM template.
Unfortunately, you cannot transfer an entire Azure Synapse Analytics workspace to another Azure Synapse Analytics workspace/subscription.
I would suggest you to vote up an idea submitted by another Azure customer.
https://feedback.azure.com/forums/307516-azure-synapse-analytics/suggestions/36256231-enable-support-for-cross-subscription-restore
https://feedback.azure.com/forums/307516-azure-synapse-analytics/suggestions/40528870-connectivity-to-code-repositories-similar-to-data
All of the feedback you share in these forums will be monitored and reviewed by the Microsoft engineering teams responsible for building Azure.
Azure SQL scripts, Notebook can be imported/exported from Azure Synapse Studio.
For pipelines, linked services you can download the support files and copy the JSON template and create a new pipeline using the same JSON template.
You can also use Azure PowerShell modules for to import and export.
For more details, refer Az.Synapse

How to Connect ADLS Gen-1 with Azure ML Studio

Want to connect ADLS Gen-1 with AzureML Studio.
I try to find out some solution but could not get
Direct method:
Currently, Azure Data Lake Store is not a supported source.
I would suggest you to vote up an idea submitted by another Azure customer.
https://feedback.azure.com/forums/327234-data-lake/suggestions/15008490-adl-store-connector-for-ml-studio
All of the feedback you share in these forums will be monitored and reviewed by the Microsoft engineering teams responsible for building Azure.
By using the Import Data module, you can access data from one of several online data sources while your experiment is running:
• A Web URL using HTTP
• Hadoop using HiveQL
• Azure blob storage
• Azure table
• Azure SQL database or SQL Server on Azure VM
• On-premises SQL Server database
• A data feed provider, OData currently
• Azure Cosmos DB
For more details, refer “Supported data types in Azure ML studio”.
In-direct method:
Azure Data Lake Analytics can also be used to write data out to Azure Blob Store, and so you can use that as an approach to process the data in U-SQL and then stage it for Azure Machine Learning to process it from Blob store. When Azure ML supports Data Lake store, then you can switch that over.
For more details, refer "How to use ADLS as an input data set for Azure ML Studio".
Hope this helps.

Azure Gov Cloud and Azure Functions trigger on Storage

I have hard time with Azure Functions on Azure Government. I need to create a C# trigger bases process on Azure Storage. The goal is to automate the process of the loading the files into Azure SQL DB when a file is dropped into Azure Storage.
Since Azure Functions in Azure Government are not fully comparable to Azure Function in regular Azure and not all UIs are the same, I can't deploy the function to trigger on a storage file.
I was able to build the process in regular Azure Cloud following instructions from https://github.com/yorek/AzureFunctionUploadToSQL but since Azure Government is missing the UI for Azure Functions I'm having hard time to replicating the process in Azure Government.
Portal UI support is not yet available in Azure Government, but it is coming soon. Additionally, Azure Government currently supports "App Service plan" ("Consumption plan" coming soon).
In the meantime, you can do everything you need. First, provision your Azure Function in Azure Gov via the Azure CLI by following this Quickstart example for Functions on Azure Gov. That same link also shows you how you can use Visual Studio to set up your triggers (in your case, a Blob trigger).
Once complete, deploy your Function to Azure Gov with Visual Studio.