I have some Rabbitmq server issues installed in Centos 7 when starting and stopping it. The server works well when connected to internet. It only takes 1 or 2 seconds to start it. But when disconnected to internet, it takes about 3~5 minutes to either start or stop it. There seems nothing wrong on the access log.
I also checked what to set up in the configuration file. But I couldn't find any settings about starting server. Can you guys give some help for me?
Related
I would like to understand the logs of a RabbitMQ error. Can anyone help me decipher the following error message:
The interesting thing is that in my case, I have a server running with RabbitMQ 3.8.8 running and erlang 23 behind it, but suddenly after a long period running this crash occurred.
Maybe it's related to port 25672 connectivity?
The error repeated for a few hours until we noticed and then the whole server was restarted, this fact made the rabbitmq run again and everything went well.
So the question is, what does this error mean, that port 25672 doesn't respond?
Dears,
I am facing a problem with my weblogic 12c Server. I have 3 managed nodes and one admin server. I have configured an Oracle Datasource. Whenever any network glitches or database downtime happens, weblogic server also goes down completely. Is there any configuration which I can make to hold on the server until everything comes back.
Servers are trying to restart automatically.But it is in STARTING state. I have to again start all servers and it is taking time. Your suggestions will help me a lot. Thanks in advance.
If you don't want your servers to restart automatically, the server failure action can probably be tweaked:
{managed_server} -> Configuration -> Overload -> Failure Action
However, probably you should be more interested in exploring if you are missing connection timeout config for your datasource.
I need to run rabbitmqctl rotate_logs on a rabbitmq log file that is over 80gb's in size. When I tried to run this the first time it froze rabbit and no messages could be received. The freeze lasted 20 mins before I had to kill the command and restart the rabbit server.
This is a production server and completing this in a small amount of time without losing messages or killing the broker would be optimal.
Would it be possible to shut down the service and move the current log file to another location and restart the service and then run the rotate_logs command?
I'm fairly new to rabbitmq and I am not sure what the best way to handle this would be.
This is installed on a windows 2008 server as a service for a heavy traffic production site (However the message queue has a small load and only affects the administrative side of things).
Any help or insight would be appreciated.
I ran into a similar situation, but with only about 4GB of log file instead of 80.
the workaround I used was pretty much what you suggested... stop the service, move the log file and restart the service as quickly as possible.
for me, specifically, instead of moving the file while the service was stopped i just renamed it. i also wrote a commandline script to do the work for me.
this allowed me to stop the service, rename the file and restart the service in a matter of seconds.
once the service was back up and running, i was free to move / rename / whatever the large log file as needed.
Can anybody suggest how to update a web-service /wcf service holding live transactions without any downtime?
Please don't suggest updating at off peak hours, because our service runs 24x7 and transactions may run any period of time.
So what is the best way to update such service with some changes so that current transactions don't get affected?
I use load balancing to prevent errors and keep the site up during updates. For this to work you need at least two servers behind a load balancer that sends traffic to whichever server is up. My procedure for updating my sites is:
Tell "Server A" to start serving an error page at the URL that the load balancer pings. This tells the load balancer to stop sending traffic to this server.
Wait 30 seconds or so until traffic stops hitting the server.
Update the code on this server.
Tell the server to stop serving an error page at the ping URL.
Wait until that server is getting traffic again.
Repeat steps 1-5 with "Server B".
From my experience... i use a Git server and a local machine to edit the code to make sure it all works, and then push to the server, then pull on the live site, and then it has 0 down time....
Without 0 seconds downtime, It is not possible.
Least you can do is Update the web-services, test that thoroughly at say staging server. Verify and be sure updates are working perfect.
Update the services- part (a xyz.war file in my case) and re-start the server for getting this into effect.
If there is any DB or other changes involved in this process, i make a script of that and test the same at staging server before going live.
Just one suggestion: Take proper backup of current live server for the otherwise case.
I have two rails app which are running on linode. OS is ubuntu, nginx server. The subdomain instance giving problem. It is getting down just after 1 day. On restarting the server, it is working fine.
The error log says- "*1 upstream timed out (110: Connection timed out) while reading response header from upstream,".
I googled for the problem and found that increasing proxy_read_timeout value will solve the problem. But I am unable to find the reason.
Is there an issue of over utilization of resources? I have 24 GB of storage and 512 MB of RAM as shown in linode manager. I have 10 cron jobs in total (5 in each app). They all start at the same time. Can that be the issue?
Please tell me the reason and solution for it.
It definitely sounds like a resource issue... Or perhaps something else is killing / hogging your app. Generally the upstream request is a request from the web server to the app server, so if your app is doing something wonky, this would cause the timeout to occur. I'm not sure what the default timeout is, but I'm guessing it's rather short. Increasing the timeout would at least buy you time to look at the system resources the process stack to try to figure out what's going on.