postgres daily query analysis tool [closed] - optimization

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 2 years ago.
Improve this question
does anyone know of a postgres view / function / tool that could report on the slowest and most often used slower queries? i think this would be so useful for all sysadmins. thanks!

PostgreSQL 8.4 comes with some addon modules that will help you with detailed analysis of this. They don't actually create reports, but they get you access to more data.
You can also look at pgfouine (http://pgfouine.projects.postgresql.org/). It generates nice-looking HTML reports.

I wrote such a tool some time ago. You can get it here (file: analyze.pgsql.logs.pl)
There are (as of now) no real docs.

pgBadger is also good log analyzer tool that you can use.
Link: https://github.com/dalibo/pgbadger/downloads

Related

Is there a way to look at the back-end processing of PostgreSQL? [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 2 years ago.
Improve this question
I am new to StackOverflow so please do correct me if I need to provide any more information.
I am trying to integrate an anomaly detection into the PostgreSQL database system by plugging into its backend.
I would like to know if there is any place where I can find extensive back-end coding or integration examples. I am looking at papers regarding this topic and, so far, I have found a few which talk about the methods that the queries are classified and used for anomaly detection.
If you do know about any websites which might help me, please do provide links to the sites.
Thanks!!
Look for "hooks" in the PostgreSQL source tree. Studying the source of the auto_explain and pg_stat_statements contrib modules, which track query execution, will show you the way.
The source is its own book: it is well-documented and interspersed with README files that explain the design.

What are the spell correct api's available? [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 5 years ago.
Improve this question
I wanted to know the spell correct api's available for commercial/non commercial usage other than google/bing.
First of all you can write your own spell corrector with this tutorial. In addition there are some Python packages that may help you with that, such as TextBlob (which I highly recommend). Another option is Gingerit which Iv'e never tried but looks promising. Another DIY spell correct tutorial might interest you as well.
https://www.gigablast.com/spellcheckapi.html
I just launched this, so it's still beta, but it's not bad. It has a dictionary of over 600,000,000 entries covering most non-Asian languages.

Open source Test case management tool [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 5 years ago.
Improve this question
I am looking for an open source test case management tool.
Is there anything which can take csv / xml / xls files and create test cases for me.
Recently found Nitrate\Kiwi, which is written in Python as Django app, may be interesting to take a look at https://github.com/kiwitcms/Kiwi
The free community edition of Klaros Testmanagement has support for importing test cases from xls and XML.
It is not open source but free to use and not time or user crippled.
Disclaimer: Being involved in the development I am biased

Pydoc equivalent for Common Lisp [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 5 years ago.
Improve this question
I've been programming in common lisp for a while now, and I like how there's so much handy documentation on the language online; the problem is that I'm often offline and can't access it when I need it most.
Is there a PyDoc equivalent for common lisp (or even just man pages for the language) that I can download and use in the shell?
Cheers in advance.
You can download the CLHS and install it in various ways.
http://www.cliki.net/CLHS
that's an old question…
edit: as referenced on the Cookbook, we can read the HyperSpecs offline with either Dash (MacOS), Zeal (GNU/Linux) or Velocity (Windows).
we could ask or add it on devdocs: https://devdocs.io/
and take the data of the CL Ultra Spec: website, data
and of course browse the built-in documentation with Emacs (C-h, see the menu).

Does autotrader.com (vehicle marketplace) offer an API for posting vehicles? [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 4 years ago.
Improve this question
Does anybody know if autotrader.com offers an API or something that would help with mass postings of vehicles?
Or does anybody have any idea of what to use to create something like this? I was thinking of maybe a mouse location and click over a browser window type of thing.
AutoTrader provides a bulk upload feature through a file feed process. The file runs through a set of processes to associate it with the proper listing tier (Premium, Feature or Standard) and in addition normalizes the information across vehicle make and models. This process runs several times daily and is being migrated to a near real-time solution for quicker add or updates.
I can find no documentation on it but I've come across this link which seems to provide a json response.
https://www.autotrader.com/rest/searchresults/sunset/base