Parsing data to tables in excel - vba

I need to scrape the data from this page into a table format. To give you an idea, Here is what it looks like :
I want to see there is a way to bring this data into a table format using VBA script in excel. I can write a VBA code to do this by trimming the information into small pieces and placing it in particular cells.
I believe that's not the most efficient way and something tells me there is probably a better method out there that I just don't know off yet. Hence, I am reaching out here.
I wanted to see if you guys know of a method to do this better than trimming and pasting.

The data is in JSON format. You have to parse it and then decide in what way you want to present the data in Excel.
Here it is informacion about that:

Related

UiPath pdf table scraping into a DataTable type

Can I somehow import a table from pdf to a UiPath DataTable?
I can do it via loading it to a string array and after that splitting it. But I hope so there is a better and more safe solution to get a table from PDF.
The easiest way is to use the Read PDF Text activity.
As it's often changing, it does really make sense to give you a tutorial here about how you add it to UiPath and what parameters you currently have. Overall all info you can find on the official detailed tutorial.
Basically this activity gives you the easy way of extracting data and managing it.
If the result is not okay you will need to switch to the OCR activity with nearly the same name. This one reads the data visually. But from what you gave here, I would not recommend you that.
There are many other activities out there, like extracting it into an Excel Workbook. So simply try out what you need.

Read Excel Files from External Tables

I am tasked to create a template that will be Filled up by Business Users with Employee Information, then our program will load this into the Database using External Tables.
However, our Business Users constantly change the template by adding, removing or reordering fields.
I am convinced to use XLSX instead of CSV so that I can lock the Column Headers so they cannot remove, add and reorder the columns.
However, When i query the External Table, it shows Non-ASCII Characters when reading XLSX because its in Binary.
How can i do either of the following?
Effectively Read Excel Files from External Tables
Lock the Headers of CSV Files?
What you have here is a political problem, but you are looking for a technical fix. Not a good fit.
The problem comes in two halves:
Somebody decided it was a good idea to collect user input in a spreadsheet, which it is generally not.
Users are fiddling with the input format, which they should not.
Fixes are:
Strictly enforce the data structure. Reject any CSV which doesn't natch and make the users edit them. They will quickly tire of tweaking the spreadsheets when they realise they're just creating more work for themselves. But they will also get resentful, so consider ...
Building a data input screen. It's pretty simple to knock up a spreadsheet-like grid UI. You don't need anything complicated in Java: Oracle's Apex is intended for exactly this sort of thing. Find out more.
However, if you are stuck with Excel as a UI I suggest you have a look at Anton Scheffer's excellent PLSQL as_read_xlsx package on the Amis site. Check it out. You'll probably need to replace your external table with a view over a table (perhaps pipelined) function.

Get PDF form field description with VBScript

I'm pretty new to VB, but I've got a script working which involves taking a PDF with fillable fields, and populating a column in an xlsx with all of the field names.
I would like to take this a step further, and pull the field descriptions as well, into another column. By that, I mean the description which comes up when mousing over a particular field in the form. Getting the name of the field is easy enough, but I've done some digging and am not sure if it's possible to pull the description as easily.
I've experimented with functions similar to "object.getNthFieldName(x)," but no such luck.
Not entirely sure if this is possible, but looking for any insight. Thanks!

T-SQL Pivot - Pivot on all rows programmatically

Situation
I have a report which is to be export by customers to excel format when they please that must meet certain formatting criteria.
Problem
This report is a statutory return that must be returned in a specific format. I have decided this can be achieved with the use of pivots. Instead of a list view they require a grid of totals, like this
o...
This can be done manually by writing out lots and lots of pivots in a big complex statement, but would need to be updated as soon as the column I am pivoting on has a new entry, which will be frequent.
Perfect Solution
The perfect solution for this would be for somebody to find something magical that automatically pivots on all data in one column and does so programmatically, so if entry's are added to that column, they will also be pivoted on.
I am not looking for exact code here because that's lazy and I won't learn anything. A shove in the right direction would be nice though, if this is at all possible.
You can do this with dynamic SQL (Example in this answer). Although this is frequently best achieved through an application/reporting layer rather than directly in SQL.
In your case you can embed reports in Windows Forms Applications, which gives the option of just viewing the report, or exporting in a number of formats. If the pivot needs to be done in SQL then dynamic sql is the only way to do it.

SSIS automate insert upate excel file format issue

i have to insert and update some values which is daily coming from
excel file but as everyday excel file format is different
so tell me other possible ways to automate insert update ?
Are the excel files really in different formats or does Excel just think they are different? If the columns are still in the same ordinal positions but they are being interpreted as having different data types, then yes, you can provide hints to the driver to overcome.
Otherwise, you could use C#/vb.net and query the worksheet, dump that into an dataset, write that to a variable and then shred that object but it's ugly. In fact, dealing with Excel in a programmatic fashion is always ugly and best avoided.
If your file is a different format each day then you are out of luck. That is a problem, and there is really no easy or efficient way to parse and insert/update based on that. Whatever the source of the data is, you need to ensure that it becomes consistent.
If it is a handful of formats that you can test and handle accordingly then you could always have some data flow logic inside the SSIS package, but if this isn't predetermined then you would have no way of handling these cases.
YOu deal with this by returning the file to the provider and requiring them to provide in the same fashion every day. Then your SSIS pacakge should reject the file if it is not in the correct format. While you aer at it you will have far fewer problems if they send .txt or .csv file. Excel support is exceedingly poor.