I am getting error when delete file after consume blob message in activemq. Here is my error stack trace.
Note - Downloaded successfully. Getting error on delete -
activeMQMessage.deleteFile();
Error:
entejavax.jms.JMSException: Delete file failed: 550 Delete operation failed.
at org.apache.activemq.blob.FTPBlobDownloadStrategy.deleteFile(FTPBlobDownloadStrategy.java:66)
at com.skynetwork.sync.FileSyncListener.onMessage(FileSyncListener.java:66)
at org.apache.activemq.ActiveMQMessageConsumer.dispatch(ActiveMQMessageConsumer.java:1298)
at org.apache.activemq.ActiveMQSessionExecutor.dispatch(ActiveMQSessionExecutor.java:131)
at org.apache.activemq.ActiveMQSessionExecutor.iterate(ActiveMQSessionExecutor.java:202)
at org.apache.activemq.thread.PooledTaskRunner.runTask(PooledTaskRunner.java:122)
at org.apache.activemq.thread.PooledTaskRunner$1.run(PooledTaskRunner.java:43)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Any help is appreciated.
it seems to be ftp server problem, can you verify if you can delete it with FileZilla or other FTP tool and if name on the consumer side contains blanks or special caracters. Did you have FTP side logs.
Related
I have deployed Pentaho biserver-ce and data-integration (6.1.0.1.196) successfully. It first time loading without any issue.
After I have deployed report set, that is already working on locally. This deploy done using pentaho API. After completing the deployment I can see following issue. Please advise me.
16:34:40,983 ERROR [FileHandler] Couldn't find or create CDE /widgets/sample.cdfde file
org.pentaho.platform.api.repository2.unified.UnifiedRepositoryAccessDeniedException: access denied while updating file with id "5117ec88-9572-43f2-9774-1a69a09c2b0e"
Reference number: 69b57b5c-c978-4292-bcb9-c1967ee04654
at org.pentaho.platform.repository2.unified.exception.AccessDeniedExceptionConverter.convertException(AccessDeniedExceptionConverter.java:31)
at org.pentaho.platform.repository2.unified.ExceptionLoggingDecorator.callLogThrow(ExceptionLoggingDecorator.java:506)
at org.pentaho.platform.repository2.unified.ExceptionLoggingDecorator.updateFile(ExceptionLoggingDecorator.java:433)
at pt.webdetails.cpf.repository.pentaho.unified.UnifiedRepositoryAccess.saveFile(UnifiedRepositoryAccess.java:182)
at pt.webdetails.cpf.repository.pentaho.unified.UnifiedRepositoryAccess.saveFile(UnifiedRepositoryAccess.java:163)
at pt.webdetails.cdf.dd.extapi.FileHandler$1.call(FileHandler.java:114)
at pt.webdetails.cdf.dd.extapi.FileHandler$1.call(FileHandler.java:110)
at org.pentaho.platform.engine.security.SecurityHelper.runAsSystem(SecurityHelper.java:396)
at pt.webdetails.cdf.dd.extapi.FileHandler.createBasicFileIfNotExists(FileHandler.java:110)
at pt.webdetails.cdf.dd.CdeEngine.saveAndClose(CdeEngine.java:135)
at pt.webdetails.cdf.dd.CdeEngine.ensureBasicDirs(CdeEngine.java:118)
at pt.webdetails.cdf.dd.CdeLifeCycleListener.ready(CdeLifeCycleListener.java:39)
at org.pentaho.platform.web.http.context.PentahoSystemReadyListener.contextInitialized(PentahoSystemReadyListener.java:52)
To start pentaho server, taking time and show the following error message as well.
16:03:48,064 INFO [PeriodicStatusLogger] Caution, the system is initializing. Do not shut down or restart the system at this time.
16:04:18,064 INFO [PeriodicStatusLogger] Caution, the system is initializing. Do not shut down or restart the system at this time.
16:04:29,865 ERROR [BlueprintContainerImpl] Unable to start blueprint container for bundle pdi-dataservice-server-plugin due to unresolved dependencies [(objectClass=org.pentaho.metaverse.api.ILineageClient)]
java.util.concurrent.TimeoutException
at org.apache.aries.blueprint.container.BlueprintContainerImpl$1.run(BlueprintContainerImpl.java:336)
at org.apache.aries.blueprint.utils.threading.impl.DiscardableRunnable.run(DiscardableRunnable.java:48)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
I'm deploying a ToolTwist application to a production server using FIP, and Im getting this error on Transfer Phase.
java.net.SocketTimeoutException: Read timed out
at java.net.SocketInputStream.socketRead0(Native Method)
at java.net.SocketInputStream.read(SocketInputStream.java:152)
and in the fipserver console
org.eclipse.jetty.io.EofException
at org.eclipse.jetty.http.HttpGenerator.flushBuffer(HttpGenerator.java:892)
at org.eclipse.jetty.http.AbstractGenerator.blockForOutput(AbstractGenerator.java:486)
at org.eclipse.jetty.http.AbstractGenerator.flush(AbstractGenerator.java:424)
at org.eclipse.jetty.server.HttpOutput.flush(HttpOutput.java:78)
at org.eclipse.jetty.server.HttpConnection$Output.flush(HttpConnection.java:1094)
at org.eclipse.jetty.server.HttpOutput.write(HttpOutput.java:159)
at org.eclipse.jetty.server.HttpOutput.write(HttpOutput.java:98)
at tooltwist.fip.jetty.GetFileListServlet.doGet(GetFileListServlet.java:82)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:707)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
Caused by: java.io.IOException: Broken pipe
at sun.nio.ch.FileDispatcherImpl.write0(Native Method)
at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)
at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:94)
what should be the solution for this?
This error is occuring in the first stage of the FIP file transfer, where the fipserver creates an index of the existing files on the destination server. This is done in GetFileListServlet.doGet(), which can be seen in the stack trace. It is also indicated on the client side by the message...
Indexing source...
Indexing destination...
ERROR: java.net.SocketTimeoutException Read timed out
Exception: tooltwist.fip.FipException: java.net.SocketTimeoutException: Read timed out
This indexing process involves creating a hash for each file on the destination server, which the fip client then compares with the hashes of files on the source machine. It does this to determine which files are different, and so need to be installed.
A read timeout occurs when the client is waiting too long for the FIP server to index the files on the destination machine. Indexing is normally a fairly quick process, but does involve reading all the files beneath the destination directory (e.g. in ~/server). If monsterously huge files exist within that destination directory then the scanning will take a proportionately long time to complete. If that time is too long, then the client times out and drops the connection, and the server also sees the connection was dropped and stops indexing.
The most common cause of this error is excessively large log files in ~/server/tomcat/logs. If you clean those up, the problem should go away.
I have created a project and added few pages on the jsp wiki and now i just want to add some pdf files, but it shows me the following exception/error
Apache Tomcat/6.0.18 - Error report HTTP Status 500 - type Exception reportmessage description The server encountered an internal error () that prevented it from fulfilling this request.exception javax.servlet.ServletException: java.io.IOException: Upload failure: Processing of multipart/form-data request failed. Disk quota exceeded
com.ecyrd.jspwiki.dav.WebdavServlet.service(WebdavServlet.java:174)
javax.servlet.http.HttpServlet.service(HttpServlet.java:717)
com.ecyrd.jspwiki.ui.WikiServletFilter.doFilter(WikiServletFilter.java:174)
root cause java.io.IOException: Upload failure: Processing of multipart/form-data request failed. Disk quota exceeded
com.ecyrd.jspwiki.attachment.AttachmentServlet.upload(AttachmentServlet.java:692)
com.ecyrd.jspwiki.attachment.AttachmentServlet.doPost(AttachmentServlet.java:473)
javax.servlet.http.HttpServlet.service(HttpServlet.java:637)
com.ecyrd.jspwiki.dav.WebdavServlet.service(WebdavServlet.java:167)
javax.servlet.http.HttpServlet.service(HttpServlet.java:717)
com.ecyrd.jspwiki.ui.WikiServletFilter.doFilter(WikiServletFilter.java:174)
note The full stack trace of the root cause is available in the Apache Tomcat/6.0.18 logs.Apache Tomcat/6.0.18
I even tried attaching the images but in that case also it shows the same error.
I also have not set any filters on the size or the type of the file for the attachments.
I'm running spring integration with udp inbound and outbound. I'm getting the following error, how can i solve this ?
13:37:12.346 ERROR [udpIn-ip:udp-inbound-channel-adapter][org.springframework.integration.ip.udp.UnicastReceivingChannelAdapter] Failed to map packet to message
org.springframework.integration.mapping.MessageMappingException: Incorrect length; expected 1700885373, received 17
at org.springframework.integration.ip.udp.DatagramPacketMessageMapper.toMessage(DatagramPacketMessageMapper.java:192)
at org.springframework.integration.ip.udp.UnicastReceivingChannelAdapter$1.run(UnicastReceivingChannelAdapter.java:146)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
You didn't show your configuration but it appears you have set check-length="true". This has to be the same on both sides (assuming you are sending from a Spring Integration app).
If you are receiving data from some other application; this attribute must be set to false (default).
I was stopping the Liferay portal, but few seconds after, I stopped the database (db2 quiesce, that means, that the connections are closed) and apparently, Liferay did not stopped correctly its execution.
After that, I restarted the database and liferay, but the portal does not work now. It shows this message in the browser:
HTTP Status 500 -
type Exception report
message
description The server encountered an internal error () that prevented it from fulfilling this request.
exception
javax.servlet.ServletException: Servlet execution threw an exception
com.liferay.portal.kernel.servlet.filters.invoker.InvokerFilterChain.doFilter(InvokerFilterChain.java:72)
...
root cause
java.lang.NoSuchMethodError: com.liferay.portal.util.PortalUtil.getCDNHostHttp()Ljava/lang/String;
com.liferay.portal.events.ServicePreActionExt.servicePre(ServicePreActionExt.java:937)
After looking in the logs, I found the following messages (they are edited):
SEVERE: Error waiting for multi-thread deployment of directories to completehostConfig.deployWar=Deploying web application archive {0}
java.lang.InterruptedException
at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1000)
WARN [DefaultConnectionTester:203] SQL State '08001' of Exception which occurred during a Connection test (fallback DatabaseMetaData test) implies that the database is invalid, and the pool should refill itself with fresh Connections.
com.ibm.db2.jcc.am.DisconnectNonTransientConnectionException: [jcc][t4][2030][11211][3.63.75] A communication error occurred during operations on the connection's underlying socket, socket input stream, or socket output stream. Error location: Reply.fill() - insufficient data (-1). Message: Insufficient data. ERRORCODE=-4499, SQLSTATE=08001
at com.ibm.db2.jcc.am.fd.a(fd.java:321)
WARN [DefaultConnectionTester:136] SQL State '08001' of Exception tested by statusOnException() implies that the database is invalid, and the pool should refill itself with fresh Connections.
WARN [C3P0PooledConnectionPool:708] A ConnectionTest has failed, reporting that all previously acquired Connections are likely invalid. The pool will be reset.
WARN [NewPooledConnection:486] [c3p0] A PooledConnection that has already signalled a Connection error is still in use!
WARN [NewPooledConnection:487] [c3p0] Another error has occurred [ com.ibm.db2.jcc.am.SqlNonTransientConnectionException: [jcc][t4][10335][10366][3.63.75] Invalid operation: Connection is closed. ERRORCODE=-4470, SQLSTATE=08003 ] which will not be reported to listeners!
com.ibm.db2.jcc.am.SqlNonTransientConnectionException: [jcc][t4][10335][10366][3.63.75] Invalid operation: Connection is closed. ERRORCODE=-4470, SQLSTATE=08003
WARN [BasicResourcePool:1841] com.mchange.v2.resourcepool.BasicResourcePool$AcquireTask#4fad5112 -- Acquisition Attempt Failed!!! Clearing pending acquires. While trying to acquire a needed new resource, we failed to succeed more than the maximum number of allowed acquisition attempts (3). Last acquisition attempt exception:
com.ibm.db2.jcc.am.SqlNonTransientConnectionException: DB2 SQL Error: SQLCODE=-20157, SQLSTATE=08004, SQLERRMC=FUT5MAN;QUIESCE DATABASE;;, DRIVER=3.63.75
ERROR [PortalJobStore:109] MisfireHandler: Error handling misfires: Unexpected runtime exception: null
org.quartz.JobPersistenceException: Unexpected runtime exception: null [See nested exception: java.lang.reflect.UndeclaredThrowableException]
Caused by: java.lang.reflect.UndeclaredThrowableException
at $Proxy279.prepareStatement(Unknown Source)
at org.quartz.impl.jdbcjobstore.StdJDBCDelegate.countMisfiredTriggersInState(StdJDBCDelegate.java:413)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.GeneratedMethodAccessor65.invoke(Unknown Source)
Caused by: java.sql.SQLException: Connections could not be acquired from the underlying database!
at com.mchange.v2.sql.SqlUtils.toSQLException(SqlUtils.java:106)
Caused by: com.mchange.v2.resourcepool.CannotAcquireResourceException: A ResourcePool could not acquire a resource from its primary factory or source.
at com.mchange.v2.resourcepool.BasicResourcePool.awaitAvailable(BasicResourcePool.java:1319)
Now, I see that it is almost impossible to start the current Liferay installation. However, I have the database (I made a full backup), and the lucene's data directory. How can I recreate a Liferay installation with these two things? I would like to recover some of this data in a new installation, but I do not how.
This is not the best solution, but I installed Liferay with a new database. Once it was configured, I change the database configuration in order to use the other one.
Probably, it was a problem with the ROOT deployment, but this is very weird.
I could recover all the data from the Lucene and the database.
The database is still quiesced and the Liferay user doesn't have the QUIESCE_CONNECT privilege.
Unquiesce the database and restart Liferay.
Using DB2 instance owner (if you're on Windows, any administrator):
db2 connect to DBNAME
db2 unquiesce database
db2 connect reset
Regards.