Recovery default base indexes

Only applies to versions 6.1.5 and older. From version 6.1.6 and later, default indexes are created automatically

If you lost or damage following index:

    |Index name       | Index ID              |
    | .security       |Pfq6nNXOSSmGhqd2fcxFNg |
    | .taskmanagement |E2Pwp4xxTkSc0gDhsE-vvQ |
    | alert_status    |fkqks4J1QnuqiqYmOFLpsQ |
    | audit           |cSQkDUdiSACo9WlTpc1zrw |
    | alert_error     |9jGh2ZNDRumU0NsB3jtDhA |
    | alert_past      |1UyTN1CPTpqm8eDgG9AYnw |
    | .trustedhost    |AKKfcpsATj6M4B_4VD5vIA |
    | .kibana         |cmN5W7ovQpW5kfaQ1xqf2g |
    | .scheduler_job  |9G6EEX9CSEWYfoekNcOEMQ |
    | .authconfig     |2M01Phg2T-q-rEb2rbfoVg |
    | .auth           |ypPGuDrFRu-_ep-iYkgepQ |
    | .reportscheduler|mGroDs-bQyaucfY3-smDpg |
    | .authuser       |zXotLpfeRnuzOYkTJpsTaw |
    | alert_silence   |ARTo7ZwdRL67Khw_HAIkmw |
    | .elastfilter    |TtpZrPnrRGWQlWGkTOETzw |
    | alert           |RE6EM4FfR2WTn-JsZIvm5Q |
    | .alertrules     |SzV22qrORHyY9E4kGPQOtg |

You may to recover it from default installation folder with following steps:

  1. Stop Logstash instances which load data into cluster

    systemctl stop logstash
  2. Disable shard allocation

    PUT _cluster/settings
      "persistent": {
        "cluster.routing.allocation.enable": "none"
  3. Stop indexing and perform a synced flush

    POST _flush/synced
  4. Shutdown all nodes:

    systemctl stop elasticsearch.service
  5. Copy appropriate index folder from installation folder to Elasticsearch cluster data node folder (example of .auth folder)

    cp -rf ypPGuDrFRu-_ep-iYkgepQ /var/lib/elasticsearch/nodes/0/indices/
  6. Set appropriate permission

    chown -R elasticsearch:elasticsearch /var/lib/elasticsearch/
  7. Start all Elasticsearch instance

    systemctl start elasticsearch
  8. Wait for yellow state of Elasticsearch cluster and then enable shard allocation

     PUT _cluster/settings
       "persistent": {
         "cluster.routing.allocation.enable": "all"
  9. Wait for green state of Elasticsearch cluster and then start the Logstash instances

     systemctl start logstash

Too many open files

If you have a problem with too many open files by the Elasticsearch process, modify the values in the following configuration files:

  • /etc/sysconfig/elasticsearch
  • /etc/security/limits.d/30-elasticsearch.conf
  • /usr/lib/systemd/system/elasticsearch.service

Check these three files for:

  • LimitNOFILE=65536
  • elasticsearch nofile 65537
  • MAX_OPEN_FILES=65537

Changes to service file require:

systemctl daemon-reload

And changes to limits.d require:

sysctl -p /etc/sysctl.d/90-elasticsearch.conf

The Kibana status code 500

If the login page is displayed in Kibana, but after the attempt to login, the browser displays “error: 500”, and the logs will show entries:

Error: Failed to encode cookie (sid-auth) value: Password string too short (min 32 characters required).

Generate a new server.ironsecret with the following command:

echo "server.ironsecret: \"$(</dev/urandom tr -dc _A-Z-a-z-0-9 | head -c32)\"" >> /etc/kibana/kibana.yml

Diagnostic tool

ITRS Log Analytics includes a diagnostic tool that helps solve your problem by collecting system data necessary for problem analysis by the support team.

The diagnostic tool is located in the installation directory: /usr/share/elasticsearch/utils/

Diagnostic tool collect the following information:

  • configuration files for Kibana, Elasticsearch, Alert
  • logs file for Kibana, Alert, Cerebro, Elasticsearch
  • Cluster information from Elasticsearch API

When the diagnostic tool collects data passwords and IP address are removed from the content of files.

Running the diagnostic tool

To run the diagnostic tool, you must provide three parameters: - user assigned admin role, default ‘logserver’ - user password; - URL of cluster API, default: http://localhost:9200

Example of a command:

./ $user $password http://localhost:9200

The diagnostic tool saves the results to .tar file located in the user’s home directory.

Verification steps and logs

Verification of Elasticsearch service

To verify of Elasticsearch service you can use following command:

  • Control of the Elasticsearch system service via systemd:
sysetmctl status elasticsearch


 elasticsearch.service - Logserver
   Loaded: loaded (/etc/systemd/system/elasticsearch.service; enabled; vendor preset: disabled)
   Active: active (running) since Tue 2023-11-14 15:17:16 CET; 5 days ago
 Main PID: 58816 (java)
   CGroup: /system.slice/elasticsearch.service
           └─58816 /etc/alternatives/jre/bin/java -Xshare:auto -Dopensearch.networkaddress.cache.ttl=60 -Dopensearch.networkaddress.cache.n.
  • Control of Elasticsearch instance via tcp port:
curl -XGET ''


  "name" : "node-1",
  "cluster_name" : "elasticsearch",
  "cluster_uuid" : "B5SDCaaKQU2JdJpsKy6quQ",
  "version" : {
    "distribution" : "opensearch",
    "number" : "2.8.0",
    "build_type" : "tar",
    "build_hash" : "db90a415ff2fd428b4f7b3f800a51dc229287cb4",
    "build_date" : "2023-07-28T09:54:26.952266Z",
    "build_snapshot" : false,
    "lucene_version" : "9.6.0",
    "minimum_wire_compatibility_version" : "7.10.0",
    "minimum_index_compatibility_version" : "7.0.0"
  "tagline" : "The OpenSearch Project:"
  • Control of Elasticsearch instance via log file:

      tail -f /var/log/elasticsearch/elasticsearch.log
  • other control commands via curl application:

      curl -xGET "http://localhost:9200/_cat/health?v"
      curl -XGET "http://localhost:9200/_cat/nodes?v"
      curl -XGET "http://localhost:9200/_cat/indicates"

Verification of Logstash service

To verify of Logstash service you can use following command:

  • control Logstash service via systemd:

      systemctl status logstash


    logstash.service - logstash
       Loaded: loaded (/etc/systemd/system/logstash.service; enabled; vendor preset: disabled)
       Active: active (running) since Wed 2017-07-12 10:30:55 CEST; 1 months 23 days ago
     Main PID: 87818 (java)
       CGroup: /system.slice/logstash.service
              └─87818 /usr/bin/java -XX:+UseParNewGC -XX:+UseConcMarkSweepGC
  • control Logstash service via port tcp:

      curl -XGET ''


         "host": "skywalker",
         "version": "4.5.3",
         "http_address": ""
  • control Logstash service via log file:

      tail -f /var/log/logstash/logstash-plain.log


  • dynamically update logging levels through the logging API (service restart not needed):

       curl -XPUT 'localhost:9600/_node/logging?pretty' -H 'Content-Type: application/json' -d'
           "logger.logstash.outputs.elasticsearch" : "DEBUG"
  • permanent change of logging level (service need to be restarted):

    • edit file /etc/logstash/logstash.yml and set the following parameter:

         *log.level: debug*
    • restart logstash service:

          *systemctl restart logstash*
  • checking correct syntax of configuration files:

      */usr/share/logstash/bin/logstash -tf /etc/logstash/conf.d*
  • get information about load of the Logstash:

      curl -XGET ''


  "host" : "logserver-test",                               
  "version" : "5.6.2",                                     
  "http_address" : "",                         
  "id" : "5a440edc-1298-4205-a524-68d0d212cd55",           
  "name" : "logserver-test",                               
  "jvm" : {                                                
    "pid" : 14705,                                         
    "version" : "1.8.0_161",                               
    "vm_version" : "1.8.0_161",                            
    "vm_vendor" : "Oracle Corporation",                    
    "vm_name" : "Java HotSpot(TM) 64-Bit Server VM",       
    "start_time_in_millis" : 1536146549243,                
    "mem" : {                                              
      "heap_init_in_bytes" : 268435456,                    
      "heap_max_in_bytes" : 1056309248,                    
      "non_heap_init_in_bytes" : 2555904,                  
      "non_heap_max_in_bytes" : 0                          
    "gc_collectors" : [ "ParNew", "ConcurrentMarkSweep" ]  

Verification of ITRS Log Analytics GUI service

To verify of ITRS Log Analytics GUI service you can use following command:

  • control the ITRS Log Analytics GUI service via systemd:

      systemctl status kibana


    ● kibana.service - Kibana                                                                                                         
       Loaded: loaded (/etc/systemd/system/kibana.service; disabled; vendor preset: disabled)                                         
       Active: active (running) since Mon 2018-09-10 13:13:19 CEST; 23h ago                                                           
     Main PID: 1330 (node)                                                                                                            
       CGroup: /system.slice/kibana.service                                                                                           
               └─1330 /usr/share/kibana/bin/../node/bin/node --no-warnings /usr/share/kibana/bin/../src/cli -c /etc/kibana/kibana.yml 
  • control the ITRS Log Analytics GUI via port tcp/http:

      curl -XGET ''


      <script>var hashRoute = '/app/kibana';
      var defaultRoute = '/app/kibana';
      var hash = window.location.hash;
      if (hash.length) {
        window.location = hashRoute + hash;
      } else {
        window.location = defaultRoute;
  • Control the ITRS Log Analytics GUI via log file:

      tail -f /var/log/messages

SIEM PLAN - Windows CP1250 decoding problem

If Siem Agent works on operation system which works using non Latin-script alphabet, the encoding of latter could cause dropping documents by logstash. In logstash log you can notice lines like the one below.

[2023-06-01T15:36:02,091][WARN ][logstash.codecs.json     ] Received an event that has a different character encoding than you configured. {:text=>"{\\\"timestamp\\\":\\\"2023-06-01T15:36:01.214+0000\\\",\\\"agent\\\":{\\\"id\\\":\\\"002\\\",\\\"name\\\":\\\"win10_laptop\\\"},\\\"manager\\\":{\\\"name\\\":\\\"SiemPlan.local\\\"},\\\"id\\\":\\\"1549035361.0\\\",\\\"full_log\\\":\\\"{\\\\\\\"type\\\\\\\":\\\\\\\"program\\\\\\\",\\\\\\\"ID\\\\\\\":78741874,\\\\\\\"timestamp\\\\\\\":\\\\\\\"2023/06/01 15:36:00\\\\\\\",\\\\\\\"program\\\\\\\":{\\\\\\\"format\\\\\\\":\\\\\\\"win\\\\\\\",\\\\\\\"name\\\\\\\":\\\\\\\"Skype\\x99 7.34\\\\\\\",\\\\\\\"architecture\\\\\\\":\\\\\\\"i686\\\\\\\",\\\\\\\"version\\\\\\\":\\\\\\\"7.34.102\\\\\\\",\\\\\\\"vendor\\\\\\\":\\\\\\\"Skype Technologies S.A.\\\\\\\",\\\\\\\"install_time\\\\\\\":\\\\\\\"20180212\\\\\\\",\\\\\\\"location\\\\\\\":\\\\\\\"C:\\\\\\\\\\\\\\\\Program Files (x86)\\\\\\\\\\\\\\\\Skype\\\\\\\\\\\\\\\\\\\\\\\"}}\\\",\\\"decoder\\\":{\\\"name\\\":\\\"syscollector\\\"},\\\"location\\\":\\\"syscollector\\\"}", :expected_charset=>"UTF-8"}

This is caused by default Windows encoding CP1250. You can change default encoding to UTF-8 by following this steps:

  1. Go to Language settings ../_images/08-00-01-language.png
  2. Open Administrative language settings ../_images/08-00-02-admin-language.png
  3. Click on Change system locale... button ../_images/08-00-03-region.png
  4. Tick the checkbox Use Unicode UTF-8.. ../_images/08-00-04-utf8.png
  5. To make this change active you have to reboot system.