saving soap ui results of test with xml - testing

is it possible to save results of run tests in SOAP UI, including xml's of used requests and replies?
i would like to export it and send it to customer, but when i looked i couldn't find no such option
regards,
Vercy

Look in the project file XML, it's all in there. Optionally you can copy/paste any request into a text editor.

Right-click in the XML editor of the request/response and select "Save as.." ;-)

Related

Is there a way to see all the Scenario names and the Scenario teststeps in the karate report when parent feature file is calling other feature files?

I have a parent feature file that calls other feature files. In the html report, i can see only the scenarios of my parent feature file. The scenarios of the feature files called by the parent is not coming in the report. If we expand the highlighted parts in the report then only we are able to see the called scenarios. In the xml or json report also we are seeing the same. Is there a way we can view the called scenarios data also in the xml or json report?
No, and this is not a priority for Karate. The recommendation is that you focus your primary flow in the main feature and call other features only for setup and utilities.
That said, Karate is open-source and anyone is welcome to contribute and make the reports better. Also refer: https://stackoverflow.com/a/66773839/143475
I'm not sure what specifically you're looking for, but if you click the hyper link on the lines with ">>" it will display all the information for the given feature file being called.

Simplest way/tool to automate API calls and to save Json results in a file?

What would be the simplest tool/editor (ideally for Mac) to run web API queries (Stateless RESTful web API) in a loop in order to store Json results in a file ?
Very simple basically, trying just to automate the following :
- a first call to get a list of IDs
- then for each each ID, doing a call to get a few values related to this ID. Values are returned in a Json file, I would like to store them in a file (csv or excel)
To test the queries, I've used "Advanced REST client" to set a request with my authentication information header and do a few API queries tests, it works well but now I basically want to create a script to get the whole set of data which is returned and save in a file. With the idea to run this script from time to time. You can't to that with "Advanced REST client", right?
Sorry it's not (yet!) a super advanced question but any help would be greatly appreciated.
You may try Postman - definitely works on (accursed) Mac

Exporting WebCenter Content to XML

I am attempting to migrate content from an Oracle's WebCenter CMS into our organizations primary CMS. All the different parts of the page templates are separate xml snippets that get pull together and converted into html for production deployments. I am trying to find a way to export the page into xml to just get at the content. I don't need styles or js or images.
There are some built in web services and an ability to create custom ones. Is there any way to get the system to output xml or get the xml to give me a mapping of all the files so I can merge them myself?
Not sure if this will do what you want, but if you add &IsSoap=1 to the end of the URL then the request is returned in XML format. You can view the page data by using the following settings:
• IsJava
• IsSoap
• IsJson
• IsPageDebug
These may help as well. Here and here.
If this is for a Site Studio website, you should be able to turn on sitestudio section tracing, clear the server ouput, view the website page, refresh the output and it should show you details about the content items it retrieved.

Is there a public website that converts swagger json to PDF for HTML?

Has anyone made where, where you just enter your swagger URL .../swagger/docs/v1
and then the website converts it to HTML, pdf, doc or whatever in a nice readable format? I'd think that site would get a lot of traffic (hint)
I know there are some things on github you can download that will convert things, but I'd think someone has made a public site so I can save some time.
You can use this website : Swagger Editor
Copy your swagger file and in the menu, select 'Generate Client' -> 'HTML' (or Dynamic HTML')
You can also use Swagger Code Generator if you want to add this step in an automatic build flow.
Try this one: swagger2html .
The document generated by swagger-codegen is indeed not readable. This project makes a neat appearance(bootstrap css) and shows the fields in the request/response models in a straightforward manner.

Retrieving dynamic text from a website in vb.net (VS2008)

I want to be able to retrieve dynamic data from a web page (share prices). I started out by retrieving the html code before I realised that as it is live data, the html code will be of little use. Although I am looking to capture specific data, all i wish to do is process a webpage that I specify which will return the text off that website and not the HTML code. Basically a copy and paste of the entire page would be great..
Any ideas would be really appreciated!
'Screen Scraping' by parsing HTML is so early 2000s...what I would do is read up on Amazon's Mechnical Turk. You can develop a queued architecture where you submit urls to this Mechnical Turk service. The service would automatically distribute these bits of work to users who would then do the dirty task of copying and pasting out the valuable stock quote information you require. Users around the world would anxiously await delivery of the next URL to their Mechanical Turk inbox...pinning for the opportunity to copy/paste out another share price for your application. Sure, it might take a few minutes to update your prices, but hey, they would be HAND parsed by REAL people around the globe! Just think of the possibilities!
Well, the HTML contains the text of the website, so you "just" need to parse the HTML.
EDIT: If the data is not in the HTML but loaded dynamically, the situation is different. As I can see, you have two options:
Find out how the data is loaded (i.e. read the JavaScript on the page). If it is updated via some web service, you could query the same web service in your program.
Use a web browser to get the data and then get the dynamic HTML tree of the page. Maybe the WPF Webbrowser control can help you with this, but I'm not sure since I've never done this myself.
Is it possible to find this same data provided in a ready-to-consume format rather than scraping HTML for it? It seems like there's probably public web-services for stock quotes.
For example: A quick search for "Stock price webservice" turned up http://www.webservicex.net/stockquote.asmx; an ASMX web-service that is easy to consume in .NET.
In your Visual Studio project you should be add a reference to this service via the "Add Web Reference" command; the dialog you're given varies depending on whether your project is targeting for .NET 2.0 or .NET 3.0/3.5.
I added a reference to the service named StockPriceProxy:
Public Function GetQuote(ByVal symbol As String) As String
Using quoteService As New StockPriceProxy.StockQuote
return quoteService.GetQuote(symbol)
End Using
End Function