I'm working on a search engine add-on.
Is it possible to add my search engine add-on into shopify frontpage?
I have research http://www.searchifyapp.com, how they can customize shopify search page?
I am not familiar with searchifyapp.com, so I won't tell you how it works, but I can tell you how it could be done.
If you want to have the shop's data indexed on your search server, then you will want to import the data from shopify on installation and use webhooks for updates. The Syncing with a Store Shopify Wiki page explains how this is done.
You can use ScriptTags to inject javascript into the storefront. Then the javascript can find and enhance/modify the search field/form (e.g. for autocompletion/suggestions as they type, or modify the url for the search results page).
If you want to custom search results (e.g. from your own search server), then you could create an Application Proxy to serve results from your own web application.
Related
I've been developing an app that users can create a profile and it can be accessed from a URL that contains their username as a slug: https://myexample.com/username
The app is working with an API and every time the user access to the above URL, by using AsyncData from Nuxt. I'm able to get the username and do an API request to get all the information from the user and display it properly.
My concerns are, by not having a static URL to access the users' profiles, is the URL ranking on google?
My analysis is that by loading the URL only with a dynamic slug, Google will not be able to notice the existence of all the users that have a URL, but I'm not sure if my thinking is correct.
My goal is to let the users find their profiles online using Nuxt SEO advantages but I'm not sure if I'm using the correct approach.
Any feedback on this will help a lot, thanks in advance.
Google will only index pages it can find when crawling your site. If no links to the users pages exist anywhere on the site, Google has no knowledge of them and they won't be indexed.
That said, you can create a sitemap file and submit this to Google, so that it has a list of all the pages you would like it to index. This way, no internal links are required. Manually creating a sitemap for websites with a large number of dynamic pages can be tedious, however there are usually tools available to automate this depending on your setup.
EDIT
As you tagged this question with Nuxt, you could take a look at #nuxtjs/sitemap.
Requirement: From my web page user has the ability to add websites and fetch a collection of web pages having the input search term.
We have been using CSE V1 API to extract web pages based on a search term on the desired websites by making use of cref parameter. Since Google stopped supporting Link CSE feature (https://customsearch.googleblog.com/2017/02/refocusing-and-looking-forward-on.html), how do I implement the dynamic site search feature?
Thanks in Advance,
Raghu
On further analysis, custom site search is possible with the siteSearch parameter.
Ref: https://developers.google.com/custom-search/json-api/v1/reference/cse/list#siteSearch.
This parameter accepts only one website as the value, so for multiple website searches, invoke the Google CSE API multiple times for each website.
While creating the search engine, select the option "Search the entire web but emphasise included sites" so that you can hook websites while invoking the CSE API.
To enable siteSearch parameter of API
Hope it helps.
-Raghu
I am creating a kind of BOT. I would like to add web search functionality to my bot. i.e. I would like my BOT to search the web for user if your says so.
To implement that I am thinking of using GOOGLE search engine. But on overviewPage of its Custom Search API, it says
Create custom search engines that search across a specified collection
of sites or pages.
Is there any way that this specified collection defaults to whole web. or is there another way round this.
I'm building a Shopify's app which suppose to add Schema.org rich snippets according to data which is remotely stored on my server. In order that Google's crawler will actually analyze this snippets, they must be loaded during the page load, and not dynamically by some JavaScript.
Any suggestions if / how can I do something like that?
Take a look at modifying an existing theme on their documentation page.
You could also try assets, based on this answer.
I have created widgets for my website(xyz.com), which can be embedded in different websites. Let's say I embed a widget which is a photo album, in another website, abc.com. The content is residing on xyz.com but is pulled via Javascript into abc.com.
Will the content generated by the widgets (Javascript) on abc.com will be indexed by search engines?
Google will not index anything that is not visible when a page is loaded with JavaScript disabled.
There is more information in this similar question:
google indexing text retrieved by ajax or javascript after page load
Also, you can test what Googlebot 'sees' by using the "Fetch as Googlebot" feature of Google Webmaster Tools.
If you want Google to index your Ajax, you can read Google's recommendations here:
http://code.google.com/web/ajaxcrawling/docs/getting-started.html
If you follow Google's scheme for Making AJAX Applications Crawlable, then Google will index content that's generated with Javascript. So will Bing and Yandex.
Implementing this scheme is somewhat involved which is why there are companies that provide it as a service that plugs in at webserver level. (I work for one of these: https://ajaxsnapshots.com