I have one query regarding pentaho cube:
Lets assume a cube (say sample ) with schema name as schema1 and cube name as cube1. I renamed
that schema and cube to schema2 and cube 2 respectively and published on schema workbench.
As my analyzer was earlier pointing to schema1 and cube1.. so inorder to make it functional, i also updated catalog entry to schema2 and cube entry to cube2.
Now when i try to run analyzer it is showing following error in log:
com.pentaho.analyzer.service.e: Unable to find catalog: Sample
at com.pentaho.analyzer.service.impl.OlapConnectionManagerImpl.createConnection(SourceFile:79)
at com.pentaho.analyzer.service.impl.c.getConnection(SourceFile:31)
at com.pentaho.analyzer.service.impl.OlapMetaDataManager.getConnection(SourceFile:49)
Note: Well i have also checked the entry of corresponding catalog tag in datasource.xml under olap folder..
so i my question is what else i need to configure inorder to make analyzer functional?
I have repeated the same exercise.. now its working.. i simply refreshed pentaho.. may be earlier i forgot to refresh pentaho.
Related
I am using SQL Server 2012. I want to track the schema changes made in a database. For example, when a new column is added to a table (weather from designer or script), a script should be generated like 'alter table tbl1 add col1 int' and so on.
I got to the Schema Changes History Report but it doesn't provide enough information required as it only shows the table changed, change type and date and time etc but does not provide the script which was generated to make that change.
Well I got to the solution via this link. It is too easy to enable
https://www.mssqltips.com/sqlservertip/1723/auto-generate-change-scripts-in-sql-server-management-studio-ssms-for-tables/
I have a mondrian cub xml schema file. I have used it a a basis for an Analysis DataSource. I can query the thing file with Saiku, and everthing works fine.
However, when I use olap4j and try to connect to it via the xmla connetor, it does not show up in the list of cubes. Other cubes show up and can be queried. I set the EnableXmla parameter to true.
Code:
NamedList<Catalog> c = oConnection.getOlapCatalogs();
NamedList<Schema> l = oConnection.getOlapSchemas();
Schema schema = oConnection.getOlapSchema("myCube");
NamedList<Cube> cubes = schema.getCubes();
When I print out the list of Catalogs, I can see it, but only one Schema and only one Cube are listed when I print them out. (Not the one I want of course)
I have refreshed and rebooted the server. Is there anything else I need to do?
We have a SQL server database that is very dynamic and is always creating new and dropping existing tables from a custom schema called 'temp' (we have a dbo schema and a temp schema). We also use SSDT to maintain and monitor changes in our schema but we are unable to use the update feature on a schema comparison because if a new table is created (say temp.MyTable) after the schema comparison is made and before the updated is attempted, SSDT invalidates the schema comparison because something has changed. At the moment, our only solution to this is to run the schema comparisons around midnight when system activity is practically non-existent but is not ideal for the person who has to do the schema comparison.
My question is, is there a way we can exlude tables from the schema comparison that are apart of the 'temp.' schema?
How are you doing the deployment? as I test I used sqlpackage.exe to publish a dacpac and sat there constantly creating new tables and it deployed without complaining.
However, there are a couple of things you can do, the first is to stop getting the deployment to stop when drift is detected:
/p:BlockWhenDriftDetected=False
This is set to true by default.
The second thing is to ignore the temp schema, but I don't think this will help unless you also stop the drift but you might want to use this filter to stop all changes to the temp schema:
http://agilesqlclub.codeplex.com/
Ed
I've never touched PervasiveSql before and now I have a bunch of .ddf and .Btr files. I read that all I had to do was create a new database in the control center and point to the folder that contains these files.
When I do this and look at the database there is nothing in it. Since I am new to Pervasive, I'm more than likely sure that I'm doing something wrong.
EDIT: Added a screen shot after running command prompt
To create a database name in the PCC, you need to connect to the engine then right click the engine name and select New then Database. Once you do that, the following dialog should be displayed:
Enter the database name, and path. The path being where the DDFs are located. In most cases the default options are sufficient.
A longer process is documented at http://docs.pervasive.com/products/database/psqlv11/wwhelp/wwhimpl/js/html/wwhelp.htm#href=uguide/using.02.5.html.
If you pointed to a directory that had DDF files (FILE.DDF, FIELD.DDF,and INDEX.DDF) when you created the database name, you should see tables listed.
If you pointed to a directory that does not have DDF files, the database will still be created but will have no tables defined. You'll either need to get DDFs from the vendor or create the table entries using CREATE TABLE (with IN DICTIONARY clauses) or use DDF BUilder to add table entries.
Based on your screen shot, you only have 10 records in FILE.DDF. This is not enough. There are minimum system tables required (X$FILE, X$FIELD, X$INDEX, and a few others). It appears your DDFs are not a valid set. Contact the client / vendor that provided the DDFs and ask for a set that include all of the table definitions.
Once you have tables listed in your Database Name, you can use ODBC to access the data.
Sorry if the question isn't phrased very well but I'm new to SSAS and don't know the correct terms.
I have changed the name of a table and its columns. I am using said table as a dimension for my cube, so now the cube won't process. Presumably I need to make updates in the analysis server to reflect changes to the source database?
I have no idea where to start - any help gratefully received.
Thanks
Phil
Before going into the details of how to amend the cube, have you considered creating a view with the same name as the old table which maps the new column names to the old?
The cube processing process should pick this up transparently.
EDIT
There are quite a lot of variations on how to amend SSAS - it depends on your local set-up.
If your cube definition is held in source control (which it should ideally be), you need to check the cube definition out and amend it from there.
If your definition exists only on the server you need to open it from the server:
Open the Business Intelligence
Development Studio (BIDS) -
typically on the Windows start menu
under Programs > Microsoft SQL
Server 2005.
Go to File > Open > Analysis Services Database
Select your server/database and click OK.
Once you have the project open in BIDS, you can amend the Data Source View to switch to the new table.
These instructions are based on the principle that it's going to be easier to alias the new table to look like the old in the DSV, since this means fewer changes within the cube definition.
Open the Data Source View from the Solution Explorer - there should be only one.
Locate the table you need to change in the DSV
Right-click on the table and select Replace Table > With New Named Query
Replace the existing query with a query from the new table with the new columns aliased with the new names:
SELECT ~new column name~ AS ~old column name~
FROM ~new_table~
Once the new query has been set, deploy the changes:
If you use source control, check in and deploy the project to the target server.
If you opened the cube definition from the server, select File > Save All
Finally, re-process the cube.