I am connected to LDAP and trying to import one ldif file under one identity. While Importing the ldif when I am clicking on "finish", I am getting the error as "error while Importing ldif, it violated the schema because it contains attribute "xyz" which is not allowed by any of the object classes in the entry".
Kindly help me on how to resolve this error.
Thanks in advance.
Related
I'm trying to connect dbt to BigQuery in vscode. For that I extracted a bigquery keyfile json that I put into the root directory of my dbt project.
I then created a profiles.yml file that looks as follows:
my-bigquery-db:
target: dev
outputs:
dev:
type: bigquery
method: service-account
project: civil-parsec-350114
dataset: dbt_dataset
threads: 1
keyfile: bigquery.json
Database Error
Runtime Error
dbt encountered an error while trying to read your profiles.yml file.
Could not automatically determine credentials. Please set GOOGLE_APPLICATION_CREDENTIALS or explicitly create credentials and re-run the application. For more information, please see https://cloud.google.com/docs/authentication/getting-started
When I put an empty projects.yml file I get the same error, so I'm not even sure if that file is loaded at all. How can I best debug this? What could be the problem?
I'm trying to install sql developer on windows7 but I got this message "unable to set the current directory based on the module file name". how to fix this error?
When I am trying to import .DMP file via SQL developer I am getting this error
Exception: ORA-31640: unable to open dump file "/home/oracle/Desktop/dump/vahe.DMP" for read
dump directory and vahe.dmp file have read and write permission.
I use Database App Development VM.
how can I fix this issue ?
Thanks.
Well I found the problem. Actually I had type error. I have typo "vahe.DMP" instead of "vahe.dmp"(in lower case ). I think error message is not good one, because it should clearly say that file does not exist instead of saying "unable to open dump file '' for read" (IMHO)
Thanks everybody who tried to help me.
I was having the same error while importing DMP file shared from colleague.
error “ora-31640 unable to open dump file for read”
By creating new user with same name and password which was used while creating DMP file, and used this user for connect and importing, it resolved this error.
I was importing the data using "Data Pump Import Wizard" in oracle 11g R2 server.
Thank you for looking into this. I am fairly new to SSIS and got stuck. I have googled a lot of resources but dont see anybody else having the same issue.
I have created an SSIS package which downloads a 'abc.zip' file from a website. I am unzipping the file using 'Execute Process Task', the unzipped file is in 'abc.DBF' format. I am importing the data from the 'abc.dbf' into sql server 2008R2. The package runs fine and exactly as intended when I run it in BIDS without config file, but when I create a config file for the package it fails at the Connection at 'DBF Source' in the Data Flow Task with the errors mentioned below:
[DBF Source 1] Error: SSIS Error Code DTS_E_CANNOTACQUIRECONNECTIONFROMCONNECTIONMANAGER. The AcquireConnection method call to the connection manager "DBF Connection" failed with error code 0xC0202009. There may be error messages posted before this with more information on why the AcquireConnection method call failed.
[SSIS.Pipeline] Error: component "DBF Source" (1) failed validation and returned error code 0xC020801C.
[SSIS.Pipeline] Error: One or more component failed validation.
Error: There were errors during task validation.
[Connection manager "DBF Connection"] Error: SSIS Error Code DTS_E_OLEDBERROR. An OLE DB error has occurred. Error code: 0x80040E21.
An OLE DB record is available. Source: "Microsoft OLE DB Service Components" Hresult: 0x80040E21 Description: "Multiple-step OLE DB operation generated errors. Check each OLE DB status value, if available. No work was done.".
I have configured the 'DBF Connection' as suggested in http://www.midnightdba.com/Jen/2010/08/tip-connect-to-dbase-files-in-ssis/ and works find without config.
I have not changed anything but just created a config file and I am using it on the same machine, with the same settings. Any help is appreciated. Thanks again.
Apologize for not being able to answer this earlier and thank each and everyone of you for taking time to look into this issue. I was able to finally get the package working by using the UNC path in my package. I saw that my package was working fine when run in the test environment but was failing when I created a configuration file. So I decided not to change anything in my Config file and the only way I thought I could do this was to use a UNC path (for the connection strings) which would allow me run the package with configuration file but without modifying anything in configuration file in production.
I have this error and i don't understand what i have to do.
I have tried all day but it doesn't work...
Error messages:
The schema script 'BOOKING_2.sch' could not be propagated to the subscriber. (Source: MSSQL_REPL, Error number: MSSQL_REPL-2147201001)
Get help: ...
The process could not read file 'C:\Program Files\Microsoft SQL Server\MSSQL11.MSSQLSERVER2012\MSSQL\repldata\unc\K-DEV$MSSQLSERVER2012_FSC_FLIPBOARD_FSC_FLIPBOARD\20130927184606\BOOKING_2.sch' due to OS error 3. (Source: MSSQL_REPL, Error number: MSSQL_REPL0)
Get help: ....
Impossibile trovare il percorso specificato.
(Source: MSSQL_REPL, Error number: MSSQL_REPL3)
Get help: ...
You need to specify a shared path and not a local path to use as your "remote" publisher. Share a folder on the local server then point the replication to \\server\shareyouhave.
Reference: MSDN
Create a network share path. And specify the path for Snapshot while configuring
\serverName\shareFolder
Please click here.