I am currently using below mentioned code in MS Access vba and code is extracting main class data but not extraction sub classes data as object required error is visible during execution.My code is given below.
Const URL As String = "https://www.tmdn.org/tmview/#/tmview/results?page=1&pageSize=30&criteria=C&basicSearch=BolĂgrafo"
Dim myproduct As Selenium.WebElement
Set d = New ChromeDriver
With d
.Start "Chrome"
.Get URL
Sleep 40000
Dim element As Selenium.WebElement
Dim elements As Selenium.WebElements
Set elements = .FindElementsByCss("div[class=rt-tr-group]")
For Each element In elements
trade = element.FindElement(By.className("sc-pZopv gQWHwO")).Text
Next element
MsgBox (trade)
Your "Sleep 40000" line may not be adequate.
since internet response times can vary widely - such an absolute time mechanism is not a reliable means of assuring that the web content is "fully loaded/retrieved".
You need to look into other means of assuring that the web page is "fully loaded".
Related
I am working on an automation project. I want to do full page (scroll down) screenshot. While the Chrome default Developer Tool (Ctr+Shift+i, Ctr+Shift+p, >capture full size screenshot) does a great job. However, is it possible to pass this instruction in VBA? I am using https://github.com/longvh211/Chromium-Automation-with-CDP-for-VBA
I have tried using VBA's sendkey methods but result is not ideal as the page is long (scroll down to one more screen). Should I use the Page.captureScreenshot() method? i.e. save the output to a string, and save the string as a PNG file.
Unable to get the output as img_string is empty:
Dim objBrowser As New clsBrowser
Dim img_string As String
Dim bytes() As Byte
objBrowser.start "edge", cleanActiveSession:=True
'By default, the new window is minimized, use .show to bring it out
objBrowser.show
'Navigate and wait 'The wait method, if till argument is omitted, will by default wait until ReadyState = complete
objBrowser.navigate "https://www.livingwaters.com/movie/the-atheist-delusion/"
objBrowser.wait till:="interactive" 'only need to wait until page is interactable. Refer to definition for other options
img_string = objBrowser.jsEval("await page.captureScreenshot()")
i just want to get the final url from a redirect.
I can read this with url = GC2.Url but the end takes a long time to load completely.
how do i get it now that as soon as i call the link with my 2nd selenium instance the webite doesn't wait for it and uses "none". but as soon as i close my 2nd instance again and continue with my first instant the normal time is used again.
i have found some approaches here: Don't wait for a page to load using Selenium in Python
approximate example:
Set GC = New Selenium.ChromeDriver
Set GC2 = New Selenium.ChromeDriver
for c = 1 to 100
GC.Get link(c)
Set Elements = GC.FindElementsByCss("p.TITLE a[href]")
For Each Element In Elements
ReDim Preserve links(a) As String
links(a) = Element.Attribute("href") 'geht auch innerHTML/href usw.
a = a + 1
GC2.Get links(a) ' this here should not wait so long
redict(a) = GC2.Url
GC2.Close
Next Element
Try to work in Headless mode, it might be faster. In this mode the browser doesn't open and everything happens with the "eyes closed", everything is virtual.
VBA code to interact with specific IE window that is already open
Above is a thread to find and go to an already open instance of IE using shell applications in VBA. AFTER I found the open IE instance I am looking for, I need to query the tables from that IE page without using it's URL. The reason that I cannot use it's URL is that this IE page is a generic 'result' page that opens in a separate window after doing a search on the main website, so if I use the URL of the result page, which is: https://a836-acris.nyc.gov/DS/DocumentSearch/BBLResult, it will return an error. Are there any other methods that allow querying tables without using URL connections, like a "getElements" for tables?
K.Davis, Tim William: you are correct in your assumptions. The first part of my code/project opens up a search page: objIE.navigate "https://a836-acris.nyc.gov/DS/DocumentSearch/BBL" and through it I submit a search form. The second part (outlined above in the first paragraph) opens up a result page (pop-up). I am trying to automate the retrieving of the tables from that page. I tried using QueryTables.Add method, the way I am familiar with to connect to the data/webpage requires an URL. If I use the URL from the result page it returns an error, thus I am looking for suggestions/help on how I could otherwise connect. That said I am able to retrieve elements of the page using 'getElements' method but not able to query tables. There are other ways to connect to the data source using the QueryTables.Add method, see, https://learn.microsoft.com/en-us/office/vba/api/excel.querytables.add but I am not familiar with these other methods. Hope this clarifies a bit.
I haven't experienced a problem with this as although you have an intermediate window the final IE window resolves to being the main IE window with focus. I was able to grab the results table with the following code using the indicated search parameters:
Option Explicit
Public Sub GetInfo()
Dim IE As New InternetExplorer
With IE
.Visible = True
.navigate "https://a836-acris.nyc.gov/DS/DocumentSearch/BBL"
While .Busy Or .readyState < 4: DoEvents: Wend
With .document
.querySelector("option[value='3']").Selected = True
.querySelector("[name=edt_block]").Value = 1
.querySelector("[name=edt_lot]").Value = "0000"
.querySelector("[name=Submit2]").Click
End With
While .Busy Or .readyState < 4: DoEvents: Wend
Dim hTable As HTMLTable
Set hTable = .document.getElementsByTagName("table")(6)
'do stuff with table
.Quit
End With
End Sub
You can copy a table via clipboard. Any tick windings appear in the right place but as empty icons.
For clipboard early bound go VBE > Tools > References > Microsoft-Forms 2.0 Object Library.
If you add a UserForm to your project, the library will get automatically added.
Dim clipboard As DataObject
Set clipboard = New DataObject
clipboard.SetText hTable.outerHTML
clipboard.PutInClipboard
ThisWorkbook.Worksheets("Sheet1").Cells(1, 1).PasteSpecial
Late bound use
Dim clipboard As Object
Set clipboard = GetObject("New:{1C3B4210-F441-11CE-B9EA-00AA006B1A69}")
I need a vbscript that could be used to copy an output from different webpages and copy it into excel sheet
Example:
Website like truecaller.Com which you can search for people by phone number.
Each number represent by unique web address ex(www.truecaller.com/au/439965324)
I need to make an excel sheet that has two columns; the 1st one is the web address and the 2nd one 8s the related name
Excel VBA is not the best for web scraping but it can get the job done.
Firstly you'll need to make sure you download the latest Internet Explorer, or at least ensure you have version 9 or above.
Secondly, you'll have to enable some references on your macros (these are analogous to imports in languages like Java). To do this, open your VBA editor, and go to Tools > References. You'll want to tick Microsoft Internet Controls and Microsoft HTML Object Library.
Now you're good to go, the following code should work for you. Not being a member of true caller, I only see "-" in the name field, but I imagine it's different if you have an account. The script I've made simply pulls out the name, number and address. I'm sure you won't have a problem with looping through your desired URLs and then placing the grabbed data where you want them.
Sub Test()
'to refer to the running copy of Internet Explorer
Dim ie As InternetExplorer
'to refer to the HTML document returned
Dim html As HTMLDocument
'open Internet Explorer in memory, and go to website
Set ie = New InternetExplorer
ie.Visible = False
ie.navigate "www.truecaller.com/au/439965324"
'Wait until IE is done loading page
Do While ie.readyState <> READYSTATE_COMPLETE
Application.StatusBar = "Trying to go to StackOverflow ..."
DoEvents
Loop
'show text of HTML document returned
Set html = ie.document
MsgBox html.DocumentElement.innerHTML
Dim element As IHTMLElement
Set element = html.getElementsByClassName("result__details")(0)
Dim Name As String
Dim Number As String
Dim Address As String
Name = element.Children(0).Children(1).innerText
Number = element.Children(1).Children(1).innerText
Address = element.Children(2).Children(1).innerText
MsgBox ("Name is " & Name & " with number " & Number & ". Address: " & Address)
'close down IE and reset status bar
Set ie = Nothing
Application.StatusBar = ""
End Sub
If you want to learn more about scraping with VBA then here's a good link:
http://www.wiseowl.co.uk/blog/s393/scraping-websites-vba.htm
I am writing a macro that will scrape my company's internal SAP site for vendor information. For several reasons I have to use VBA to do so. However, I cannot figure out why I keep getting these three errors when I attempt to scrape the page. Is it possible that this has something to do with the UAC integrity model? Or is there something wrong with my code? Is it possible for a webpage using http can be handled differently in internet explorer? I am able to go to any webpage, even other internal webpages, and can scrape each of those just fine. But when i attempt to scrape the SAP page, i get these errors. The error descriptions and when they occur are:
800706B5 - The interface is unknown (occurs when I place breakpoints before running the offending code)
80004005 - Unspecified error (occurs when I don't place any errors and just let the macro run)
80010108 - The Object invoked has disconnected from its clients. (I can't seem to get a consistent occurrence of this error, it seems to happen around the time that something in excel is so corrupted that no page will load and i have to reinstall excel)
I have absolutely no idea what is going on. The Integrity page didn't make much sense to me, and all the research I found on this talked about connecting to databases and using ADO and COM references. However I am doing everything through Internet Explorer. Here is my relevant code below:
Private Sub runTest_Click()
ie.visible = True
doScrape
End Sub
'The code to run the module
Private Sub doTest()
Dim result As String
result = PageScraper.scrapeSAPPage("<some num>")
End Sub
PageScraper Module
Public Function scrapeSAPPage(num As Long) As String
'Predefined URL that appends num onto end to navigate to specific record in SAP
Dim url As String: url = "<url here>"
Dim ie as InternetExplorer
set ie = CreateObject("internetexplorer.application")
Dim doc as HTMLDocument
ie.navigate url 'Will always sucessfully open page, regardless of SAP or other
'pauses the exection of the code until the webpage has loaded
Do
'Will always fail on next line when attempting SAP site with error
If Not ie.Busy And ie.ReadyState = 4 Then
Application.Wait (Now + TimeValue("00:00:01"))
If Not ie.Busy And ie.ReadyState = 4 Then
Exit Do
End If
End If
DoEvents
Loop
Set doc = ie.document 'After implementation of Tim Williams changes, breaks here
'Scraping code here, not relevant
End Function
I am using IE9 and Excel 2010 on a Windows 7 machine. Any help or insight you can provide would be greatly appreciated. Thank you.
I do this type of scraping frequently and have found it very difficult to make IE automation work 100% reliably with errors like those you have found. As they are often timing issues it can be very frustrating to debug as they don't appear when you step through, only during live runs To minimize the errors I do the following:
Introduce more delays; ie.busy and ie.ReadyState don't necessarily give valid answers IMMEDIATELY after an ie.navigate, so introduce a short delay after ie.navigate. For things I'm loading 1 to 2 seconds normally but anything over 500ms seems to work.
Make sure IE is in a clean state by going ie.navigate "about:blank" before going to the target url.
After that you should have a valid IE object and you'll have to look at it to see what you've got inside. Generally I avoid trying to access the entire ie.document and instead use IE.document.all.tags("x") where 'x' is a suitable thing I'm looking for such as td or a.
However after all these improvements although they have increased my success rate I still have errors at random.
My real solution has been to abandon IE and instead do my work using xmlhttp.
If you are parsing out your data using text operations on the document then it will be a no-brainer to swap over. The xmlhttp object is MUCH more reliable. and you just get the "responsetext" to access the entire html of the document.
Here is a simplified version of what I'm using in production now for scraping, it's so reliable it runs overnight generating millions of rows without error.
Public Sub Main()
Dim obj As MSXML2.ServerXMLHTTP
Dim strData As String
Dim errCount As Integer
' create an xmlhttp object - you will need to reference to the MS XML HTTP library, any version will do
' but I'm using Microsoft XML, v6.0 (c:\windows\system32\msxml6.dll)
Set obj = New MSXML2.ServerXMLHTTP
' Get the url - I set the last param to Async=true so that it returns right away then lets me wait in
' code rather than trust it, but on an internal network "false" might be better for you.
obj.Open "GET", "http://www.google.com", True
obj.send ' this line actually does the HTTP GET
' Wait for a completion up to 10 seconds
errCount = 0
While obj.readyState < 4 And errCount < 10
DoEvents
obj.waitForResponse 1 ' this is an up-to-one-second delay
errCount = errCount + 1
Wend
If obj.readyState = 4 Then ' I do these on two
If obj.Status = 200 Then ' different lines to avoid certain error cases
strData = obj.responseText
End If
End If
obj.abort ' in real code I use some on error resume next, so at this point it is possible I have a failed
' get and so best to abort it before I try again
Debug.Print strData
End Sub
Hope that helps.