Automatically verify my website's links are pointing to urls that exist? [closed] - testing

As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
Is there a tool to automatically search through my site and test all the links? I hate running across bad urls.

Xenu link sleuth is excellent (and free)

w3.org checklink

If I were you, I'd check out the W3C Link Checker.

Something like this should work: http://www.dead-links.com/
Do google searches for "404 checker" or "broken link checker"

I used Xenu's Link Sleuth in the past. It will crawl your site and tell you which links point to nowhere. It is not super fancy but it works.
http://en.wikipedia.org/wiki/Xenu%27s_Link_Sleuth
The Wikipedia page lists a whole bunch of other products.

WebHTTrack
Can take a long time to go through a large web site (I archived a 250MB website and it took approximately 2 hours - it wasn't local though) It has a log so you should be able to track 404s easily.

Also check out Google's webmaster tools.
http://www.google.com/webmasters/tools/
They give you the ability to see the 404's that GoogleBot discovers when crawling your website (along with lots and lots of other stuff).

Related

What to use now Google News API is deprecated? [closed]

As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
As part of a Project I'm working on I've been instructed to implement Google News API into a Web Application.
However, I've checked the Google News API site, and I see the following message:
Important: The Google News Search API has been officially deprecated
as of May 26, 2011. It will continue to work as per our deprecation
policy, but the number of requests you may make per day may be
limited.
I've checked SO Questions but I've not been able to find a question related to the News API.
What should I use now that Google News API is redundant?
Is it the Custom Search API?
And if so, how can I make this relevant for just
News Results for a particular query for my Web Application?
I've checked the Google News RSS, but this uses HTML in the description which won't work for my requirements as I just need the text.
Depending on your needs, you want to use their section feeds, their search feeds
http://news.google.com/news?q=apple&output=rss
or Bing News Search.
http://www.bing.com/toolbox/bingdeveloper/
I'm running into the same issue with one of my own apps. So far I've found the only non-deprecated way to access Google News data is through their RSS feeds. They have a feed for each section and also a useful search function. However, these are only for noncommercial use.
As for viable alternatives I'll be trying out these two services: Feedzilla, Daylife
Looks like you might have until the end of 2013 before they officially close it down.
http://groups.google.com/group/google-ajax-search-api/browse_thread/thread/6aaa1b3529620610/d70f8eec3684e431?lnk=gst&q=news+api#d70f8eec3684e431
Also, it sounds like they are building a replacement... but it's going to cost you.
I'd say, go to a different service. I think bing has a news API.
You might enjoy (or not) reading: http://news.ycombinator.com/item?id=1864625

What is best dojo learning material? [closed]

As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
I just want to know what is best learning tools for learning dojo for beginner.
Also is there any video tutorials for the same.
Best way to learn dojo is to try is out yourself. I would recommend following books though:
Dojo: The Definitive Guide
Mastering Dojo (From Pragmatic Bookshelf)
http://www.sitepen.com/blog/
These are the largest collection of tutorials that I know of, and cover newer features that the books don't (though the books mentioned are invaluable). You may know about them already as pretty much any Google search for Dojo turns them up.
Also the Dojo IRC is helpful and friendly for specific question.
Download the code, demo and docs from http://download.dojotoolkit.org/release-1.10.0/ and put it in a folder in your document root.
In the folder there must be the following folders
dijit/
dojo/
dojox/
util/
Open the test file in a webbrowser
http://localhost/dojo-release-1.10.0-src/dijit/tests/layout/test_SplitContainer.html
Open the file in your favorite text editor change the code and see the result.
The test files are the main resource to learn dojo.
The util folder contains some useful scripts for combining the dojo files and minifying it. This is not that useful as beginner.

Alternative to Google Custom Search [closed]

As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
I'm using Google Custom Search on a client website. They are not very happy about rival companies showing up on sponsored links on their own site. I know we can use Google Site Search but it has an annual fee. I've been looking all around for a Free/OpenSource alternative for Google CSE, but found little I can use. Anyone have any suggestions?
Check this question. What is a good search engine for embedding in a web site. IMO if the client dont want to pay for a search engine then they will have to live with the advertisements if they want a good search engine.
Google has a paid version of search. You can read about it here. We use it in our intranet.
Check out the Google JSON/AJAX Search API. It's a lightweight way of doing a query and returning pure search results that you can then display.
http://code.google.com/apis/ajaxsearch/documentation/
Search is very big business right now because it is relatively immature as an industry - similar to the OS industry many years ago. Anyone with something good is going to charge for it. The open source community will only catch up when the core concepts around search stabilize and become more widely understood (and therefore reproducible). Right now much of the basics are still trade secrets.
Short answer - if you want something even remotely as good as Google, expect to pay for it.
You can block your competitors just as with AdSense, "While AdSense allows you to filter ads by URLs, you can also filter URLs from your search results within your CSE account." - https://www.google.com/support/adsense/bin/answer.py?answer=91652
The Austrian company Mindbreeze ( http://www.mindbreeze.com/index_en.html ) has a good alternative site search for websites.
You can test it here for free: http://www.mindbreeze.com/RegisterInSite.html
opensource alternative:
http://lucene.apache.org/

What are the best wiki engine with good authorization features? [closed]

As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
What is the best wiki engine with good authorization features?
DokuWiki is a great open source PHP Wiki, clean syntax, straightforward install, simple report for read-only and private pages, plus a full ACL system, and a active community of plugin developers.
MediaWiki is amazing from a feature stand point, if you need SVG and MathML support, but I find it too hard to customize and hack on personally.
They use MoinMoin in some subjects at my uni. It has really good authorisation, controlled via simple ACLs at the top of each page. Each student submits work on pages private to them and the subject staff, so it definitely seems to work...
We use atlassian confluence it has authorization features like restricting viewing or editing of pages for groups or users. But it's a commercial product (not free)
Define "good". =)
Some would say that the original wiki script has the best authorization features in a wiki sense of the word, seeing that Ward Cunningham defined Wiki as:
The simplest online database that
could possibly work.
That said, I can share my experience with hosting Mediawiki sites and that is that I often find that I would like to be able to define privileges on a group/page and group/section kind of way. But it always depends on the application. I use UseModWiki for robowiki.net and it works excellently even though it has the most rudimentary authorization system (next after the original wiki script then).

(Apache) Error log beautifier [closed]

As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
Anyone knows of a good error log beautifier? (should be able to handle apache error logs).
Should be open source / free, preferably with a web interface
I use http://www.librelogiciel.com/software/ScanErrLog/action_Presentation It runs through the error log producing a summary webpage, or some other formats. While it's best run from a regular cron-job (and it will keep a record of what it's parsed before to save effort), it can also be run as a CGI (though the demo appears to be broken).
You can see a sample report generated using Analog and Report Magic at this address: http://www.reportmagic.org/sample/index.html. The Failure Report is simple, but is a starting point.
I use Webalizer to process my website's logs. Its web-interface might be a bit dated but it's very powerful, reliable and can handle very large logs. It's also quite easy to set up which a big plus in my book as you don't want to be worrying about the software, you just want to visualise the data.
Apache Chainsaw