What happened:
trying to export alert notifications, data-sources and dashboard through API but only getting [] with /api/alert-notifications and only receiving dashboards folder names with /api/dashboards, data-source getting all information with /api/datasources/
What you expected to happen:
expecting to get every alert notifications with there parameters and dashboards information
How to reproduce it (as minimally and precisely as possible):
fresh helm grafana installation with any dashboard, data source or alert notification
Anything else we need to know?:
tried with api key auth, user/passwiord auth, with different orgs, different permitions etc...
Environment:
Grafana version: open source 8.5.3
Data source type & version: prometheus 2.34.0/ loki 2.4.2
OS Grafana is installed on: helm installation
User OS & Browser: manjaro kde, curl/postman
Grafana plugins: json
Related
I created this dashboard by importing its ID.
Then, in order to have the necessary metrics, used this chart to install this exporter in my EKS cluster:
helm repo add prometheus-process-exporter-charts https://raw.githubusercontent.com/mumoshu/prometheus-process-exporter/master/docs
helm install --generate-name prometheus-process-exporter-charts/prometheus-process-exporter
All the prometheus-process-exporter are up and running, but the only log they have is:
2022/11/23 18:26:55 Reading metrics from /host/proc based on "/var/process-exporter/config.yml"
I was expecting to automatically have all default processes listed in the dashboard as soon as I deployed the exporter, but the dashboard still say "No data":
Do you have any ideas on why this is happening? Did I miss any step in configuring this exporter?
On a fresh install of PCF Dev, after logging in:
cf create-service p-rabbitmq standard my_rabbitmq
Showing info of service my_rabbitmq in org system / space system as admin...
name: my_rabbitmq
service: p-rabbitmq
bound apps:
tags:
plan: standard
description: RabbitMQ is a robust and scalable high-performance multi-protocol messaging broker.
documentation:
dashboard: https://rabbitmq-management.local.pcfdev.io/#/login/mu-3ea7453d-7cff-44bd-b7a7-ce0290d9b4d6-v509qlipcuhnu6relaeuat49ca/25849257198976921347999121731293969259
Showing status of last operation from service my_rabbitmq...
status: create succeeded
message:
started: 2018-01-05T01:48:17Z
updated: 2018-01-05T01:48:17Z
When I go to the management dashboard url, the browser displays the following error:
404 Not Found: Requested route ('rabbitmq-management.local.pcfdev.io')
does not exist.
How can I get the RabbitMQ Management Dashboard to be installed and respond?
The only fix I found was installing an earlier version of PCF Dev. This solved the issue.
For more details check this thread on Github.
I've downloaded Apache Apex 3.5.0 along with Malhar 3.5.0.
I've successfully started the apex client and submitted the Yahoo Finance demo example to our YARN cluster (running CDH 5.10). The cluster is running and configured properly (many Spark and MR jobs are running on it).
I see the application I submitted as RUNNING in YARN as well as in the Apex cli. However when I try to connect to the Application Master I get a 404.
org.apache.hadoop.yarn.webapp.WebAppException: /: controller for default not found
I also tried directly to connect to the appMasterTrackingUrl reported by get-app-info command, and I get the same error.
I tried a couple of apex examples, and I always get the same error.
Any idea why?
It is somewhat expected. Add "/ws/v2/stram/info" to the URL path
When you connect to the App Master you need to provide the complete URL for a REST API to invoke. There is nothing to show/return for "/" so what you are seeing is expected. What are you trying to do connecting to the App Master?
My Node Red application in IBM BlueMix is repeatedly crashing - once an hour - with no real error message other than "exited with status: 1."
How can I troubleshoot this issue?
Is there someone from IBM BlueMix support that monitors this that could take a look?
I looked at my logs and there's nothing in there that really says what's going on.
Edit per requests:
The regular log for "OUT/ERR" is scrolling so fast with HTTPD logs that I can't get it to copy/paste. Filtering to "ERR" Channel the only thing I see is below. I believe this is an error which occurs during deploy when the application restarts.
[App/0] ERR js-bson: Failed to load c++ bson extension, using pure JS version
My Node Red application is gathering data from Wink, LIFX, and other IoT services and compiles them together into a Freeboard dashboard.
Caught crash on screenshot here -- not enough cred to post images so it'll only post as a link
The zlib error was fixed in the 0.13.2 Node-RED release (that shipped 19/02/16).
If you re-stage your application is should pick up the new version of Node-RED
You can re-stage the application using the cf command line management application:
cf restage <app name>
I set up Worklight V6 Server and IWAP.
I found my worklight app console has analytics tab and there is IWAP console.
there is dashboad view, search view, search log view, geo analytics view.
then I put WL.Logger.error and WL.Analytics.log code in my app and issued these logs
but I cannot find any data on my IWAP console.
and there is following NumberFormatException in my IWAP logs. can I fix this?
[2013-06-24 18:02:35,998][DEBUG][action.search.type ] [Rattler] [worklight][7], node[M8YymIEGQbae4fbtkc2cyA], [P], s[STARTED]: Failed to execute [org.elasticsearch.action.search.SearchRequest#465651a2]
org.elasticsearch.search.SearchParseException: [worklight][7]: from[0],size[-1],sort[<custom:"worklight_data.timestamp": org.elasticsearch.index.field.data.longs.LongFieldDataType$1#79b8644>!]: Parse Failure [Failed to parse source [{"sort": {"worklight_data.timestamp": {"order": "desc"}}, "from": 0, "script_fields": {}, "facets": {}, "query": {"query_string": {"query": "worklight_data.log.message:* AND worklight_data.timestamp:[NaN TO * ]"}}, "size": 1000}]]
at org.elasticsearch.search.SearchService.parseSource(SearchService.java:566)
at org.elasticsearch.search.SearchService.createContext(SearchService.java:481)
at org.elasticsearch.search.SearchService.createContext(SearchService.java:466)
at org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:236)
at org.elasticsearch.search.action.SearchServiceTransportAction.sendExecuteQuery(SearchServiceTransportAction.java:141)
at org.elasticsearch.action.search.type.TransportSearchQueryThenFetchAction$AsyncAction.sendExecuteFirstPhase(TransportSearchQueryThenFetchAction.java:80)
at org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction.performFirstPhase(TransportSearchTypeAction.java:205)
at org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction.performFirstPhase(TransportSearchTypeAction.java:192)
at org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction$2.run(TransportSearchTypeAction.java:178)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1121)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:614)
at java.lang.Thread.run(Thread.java:779)
Caused by: java.lang.NumberFormatException: For input string: "NaN"
at java.lang.NumberFormatException.forInputString(NumberFormatException.java:76)
at java.lang.Long.parseLong(Long.java:452)
at java.lang.Long.parseLong(Long.java:494)
at org.elasticsearch.index.mapper.core.LongFieldMapper.rangeQuery(LongFieldMapper.java:176)
at org.apache.lucene.queryParser.MapperQueryParser.getRangeQuerySingle(MapperQueryParser.java:342)
at org.apache.lucene.queryParser.MapperQueryParser.getRangeQuery(MapperQueryParser.java:331)
at org.apache.lucene.queryParser.QueryParser.Term(QueryParser.java:1496)
at org.apache.lucene.queryParser.QueryParser.Clause(QueryParser.java:1319)
at org.apache.lucene.queryParser.QueryParser.Query(QueryParser.java:1275)
at org.apache.lucene.queryParser.QueryParser.TopLevelQuery(QueryParser.java:1234)
at org.apache.lucene.queryParser.QueryParser.parse(QueryParser.java:206)
at org.elasticsearch.index.query.QueryStringQueryParser.parse(QueryStringQueryParser.java:212)
at org.elasticsearch.index.query.QueryParseContext.parseInnerQuery(QueryParseContext.java:188)
Please check that the following:
1 - Are you using the developer edition? In Worklight v6, the Analytics console will not function in the developer edition. This may be changed in future releases. I have never seen that exception before, but I wonder if it occurs as a result of trying to run with the developer edition.
2 - In your initOptions.js, analytics has been set to true:
analytics : { enabled: true }
3 - In worklight.properties, you should set the queue size to 1 so that analytics are immediately seen on the console
wl.analytics.queue.size=1
Important Note The queue size should only be set to 1 in testing mode, as using 1 will not scale in production mode.
If you continue to have issues, please post more information such as the calls you are making to WL.Analytics.log and your worklight.properties for the analytics.
EDIT
If you are running Worklight 6.0 in development mode, you will need the following flag in worklight.properties:
wl.analytics.debug=true