I am working on a research project in which I need to find the in-links for approx. 170K URIs. My first thoughts were to use the Google or Yahoo APIs but then I realized that Yahoo Boss is now at Bing, which doesn't seem to support inlink queries, and that Google, deprecated its search API a while ago and replaced it with Custom Search, which doesn't seem to support inlink queries over their whole web index.
Are there any alternative solutions I missed? I am looking for a simple Web API that accepts a given URI and returns the inlinks for that URI.
Thx,
B
Update:
Google: for retrieving XML feeds of search results I need to convert a given custom search engine to Google Site Search, which is a commercial service. Even then I am only allowed to retrieved inlinks for a pre-defined set of sites, not for the whole web
Yahoo: Site Explorer unfortunately shut down and moved to Bing
Bing: in Bing's Webmaster tools app you can view the inlinks for a specific site, but you cannot query inlinks for arbitrary URIs, because it Bing Webmaster Tools doesn't provide an API yet
The SEOmoz API - If you just need to get total amount of links pointing to a URL you should be able to use the free version.
Related
I recently developed a new REST API for a company. I've created a Postman collection including some sample requests for real-world scenarios. I need to share this collection with other members of the organization.
I found a "Publish Docs" menu option for Postman collections which is pretty cool. It publishes web pages of the documentation and then displays the url for where the documentation can be accessed:
https://documenter.getpostman.com/view/1401123/RWML234Hd
One issue is that the documentation is publicly available to anyone who has the url. Meaning that anyone on the internet could potentially access the documentation. This is a corporate API and should only be accessible within the organization.
Another issue is that it appears that the published url changes every time I "Publish Docs." That's a problem b/c I need to provide a static url for Postman examples on the corporate intranet page listing API resources. My manager won't spend any money on this. I like the documentation feature but:
The documentation should have a static url
The documentation should not be accessible by anyone on the internet who gets a hold of the url
Would I need to pay license costs to satisfy the 2 requirements listed above? What would be my best free option? The default free option for me would be to save off the collection json to a shared drive on the network.
I've been asking this very question for a while now. Finally decided to implement a simple tool that generates an HTML document similar to what postman does.
hope this helps -
https://github.com/karthiks3000/postman-doc-gen
The API Documentation states that the Content API can be used to:
Edit and publish blog posts and web pages, improve SEO, and drive traffic to stores.
I'm working on migrating a site from a different platform onto BigCommerce. I was able to migrate all of the blog posts fairly easily using the API, but the documentation for the web pages resource is completely missing.
Any help here would be greatly appreciated. Thank you.
This API endpoint doesn't currently exist. It is, however, in progress. I don't have an available ETA to provide around this. This wording on the docs is misleading.
I'm building a website for a company, I build this website offline so google won't index it, with the reason that google don't see the codes yet and thinks at the official launch that we copied the text/codes from an other website.
Sadly I encounterd a problem. I need to implant the facebook social plugin, sadly does this plugin only work when the site is online. But as I said putting it online can be dangerous for future Google SEO.
Is there an other option where I can see the facebook plugin but it is not online yet or is it okay for me to just place it online already on (for example) www.example.com and later put the released product on (for example) www.released.com.
Any toughts on this problem?
Why don't you place the website online in a folder that is blocked for Googlebot via robots.txt?
That way you can test all the online elements work and not have to worry about users and search engines finding it and listing it.
An alternative could be to use the .htaccess file to limit access only to your IP address - that way, you'd be the only one to see the site live.
I want to do web search with java and I found GSA-JAPI does it work on web searching or it's used only on a website which I created for example.
The GSA-JAPI's purpose is to access and manage settings in a Google Search Appliance, which is an enterprise product Google sells. You can't use the GSA-JAPI to script any web searches, only manage a Google Search Appliance. It's used as an alternative to logging into the web based Admin Console.
The GSA-JAPI is a wrapper that serializes the XML response from the Google Search Appliance into POJOs. Additionally, You can use it to initiate searches to the Google Search Appliance. It hasn't been updated in quite some time so new elements are not exposed (i.e. Dynamic Navigation)
I'm working on redesigning an older site and I'm upgrading it to HTML5. The site currently has both Google and Yahoo site verification META tags.
I'm using HTML5 Reset as a starting template and it only has an area for Google site verification, not Yahoo. Also, the W3C Validator validates Google site verification in HTML5, but not Yahoo.
Does anyone have an opinion as to whether or not Yahoo site verification is important or useful? And does it hold any weight with SEO these days?
No.
Yahoo search is now powered by Bing.
You can remove both (Google and Bing) verification references and authenticate using other methods (DNS is best). Bing only offers meta tag and XML methods.