Connecting Lucidworks With Documentum - apache

I am building an enterprise search application in which i am using the lucidworks as a search engine and EMC Documentum as a back end technology for storing documents and meta Data.Currently i am using DQL to fire up queries and fetch data in intermediate scenario but i am looking for some other instruments to connect the two maybe third party connectors or anything else.Please Suggest me the possible ways to connect the lucidworks with documentum.

LucidWorks offers a Connector that extracts content from documentum. So you can use LucidWorks Search as well to search for documentum content.
http://www.lucidworks.com/market_app/documentum-connector-for-lucidworks/
regards
Markus

Related

API Key use w/ Tableau

We would like to use the Podio API Key to directly connect to Tableau and have the data refreshed at a cadence set in Tableau. Is this possible?
Yes, connecting to data via an API is possible and there are a couple of ways to do it:
Option #1: Web Data Connector
A WDC is a hosted web application built with JavaScript that connects to an API, converts the data to a JSON format, and passes the data to Tableau. You'll require a webserver to host your WDC and JavaScript skills to write it. Once set up, anyone in your org can just grab the link and use it in Desktop. With WDCs, since the data connection is made when the end-user is requesting the WDC, you can build in customizations for your users. (Ex: users can add filter parameters or authenticate with their own user/pass to only get what they have access to). WDC connections are extracts and can be refreshed on Tableau Server and Online. If you're using Tableau Online you'll need to use Bridge to auto-refresh.
Option #2: Hyper API
The Hyper API allows you to create, modify, and update extract (.hyper) files that you can then publish to Tableau Server/Online on a regular cadence. It's available for Python, Java, .NET, and C++ so you will need skills in one of those languages. I suggest Python as we have the most samples for it. You'll also need a server where you can run the extract refresh and publish scripts on a schedule. With the Hyper API you are creating a single extract for everyone to connect to. Once published to Tableau Server/Online, end users can just connect to this data source directly without having to do any input but this also means the connection can't be customized per user or use case.
Option #3: Use a 3rd-party connector
If building your own connector doesn't appeal to you there are also plenty of services out there you can pay for that can bring your data into Tableau. Ex: tray.io, dataddo, and skyvia are a couple I found after a quick google search.

Lucene vs. Azure Search

I am a newbie to search engines and information retrieval. Can someone explain how different is Lucene search engine compared to Azure Search.
I read the Azure Search documents and see that Azure Search supports Lucene queries as well, so is Azure Search built on top of Lucene or inherits certain features of it?
There is no proper documentation as such, can someone point me in the right direction.
Thanks in advance.
According to this Microsoft page the full text search is built on Lucene.
https://learn.microsoft.com/en-us/azure/search/search-what-is-azure-search
"The full text search engine in Azure Search is built on Apache Lucene, an industry standard in information retrieval."
Azure Search is not built on top of Apache Lucene as such, but it does support Lucene Query syntax.
https://learn.microsoft.com/en-us/rest/api/searchservice/lucene-query-syntax-in-azure-search

How to Migrate data from D2 documentum to alfresco?

I have some content in D2 Documentum. and i want to migrate all those content with meta-data in alfresco.
can any one help me.
how can i achieve this in alfresco.
Thanks in advance.
There are several tools available to handle Database/CMS migrations. If both repositories are CMIS compliant the problem becomes even easier to address. I read an interesting article about doing an Alfresco to Nuxeo migration using Apache's Chemistry and Camel offerings. Check it out.
One tool that makes migrating between ECMs very simple is Simflofy. They have connectors OOTB for CMIS compliant repositories and a very handy UI that allows easy mapping of properties. This particular tool is very fast as it was developed on a new technology stack that takes advantage multithreading and gives full audit support/reporting.
Full Disclosure: I work for a company that does several migrations a year using Simflofy.
It really depends on how many documents you need to export.
A simple answer can be: do a bulk export from Documentum and a bulk import in Alfresco.
There are several tools to export documents and metadata from Documentum. I used only a proprietary tool in my experience, but I do not think it will be difficult to find one.
For the import in Alfresco you can use the OOTB Bulk import
As you can see, you will probably need to write a small application to transform the export to the predefined format that the bulk import accepts.

SQL Metadata Search via SharePoint

Is it possible to search the metadata of SQL Databases (example: extended properties) via SharePoint? maybe using FAST?
You can use SharePoint Business Connectivity Services (BCS) to define the structure of your remote sql database and then you can crawl it with search. An intro to this is here: http://blogs.msdn.com/b/taj/archive/2010/08/24/searching-external-systems-using-sharepoint-2010-business-connectivity-services-bcs-within-throttling-limits.aspx

Semantic Application Development

hi every body
i was looking for a solution or a methodology to develop a sample
web site that utilizes a semantic database backe-end
i have used protege in previous steps and i have converted my OWL project to
mysql database backend successfully.
Any help would be appreciated
I did not see any such resource.
In general you will need a server to host your data. If your data is in relational databases you will need server like
D2R
or RDF server like
Sesame
and
Virtouso
These servers will allow accessing RDF data. Then you can use different techniques to embed this data in HTML to present it to user(Google it). There are RDF Browsers available as well.
There is a very good article on this topic here based on this publication.