8.2.1. Fichiers communs

Les composants de la solution logicielle VITAM utilisent un socle de fichiers communs.

8.2.1.1. Fichier /vitam/conf/<composant>/sysconfig/java_opts

Ce fichier définit les JVMARGS.

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
#*******************************************************************************
# Copyright French Prime minister Office/SGMAP/DINSIC/Vitam Program (2015-2022)
#
# contact.vitam@culture.gouv.fr
#
# This software is a computer program whose purpose is to implement a digital archiving back-office system managing
# high volumetry securely and efficiently.
#
# This software is governed by the CeCILL 2.1 license under French law and abiding by the rules of distribution of free
# software. You can use, modify and/ or redistribute the software under the terms of the CeCILL 2.1 license as
# circulated by CEA, CNRS and INRIA at the following URL "https://cecill.info".
#
# As a counterpart to the access to the source code and rights to copy, modify and redistribute granted by the license,
# users are provided only with a limited warranty and the software's author, the holder of the economic rights, and the
# successive licensors have only limited liability.
#
# In this respect, the user's attention is drawn to the risks associated with loading, using, modifying and/or
# developing or reproducing the software by the user in light of its specific status of free software, that may mean
# that it is complicated to manipulate, and that also therefore means that it is reserved for developers and
# experienced professionals having in-depth computer knowledge. Users are therefore encouraged to load and test the
# software's suitability as regards their requirements in conditions enabling the security of their systems and/or data
# to be ensured and, more generally, to use and operate it in the same conditions as regards security.
#
# The fact that you are presently reading this means that you have had knowledge of the CeCILL 2.1 license and that you
# accept its terms.
#*******************************************************************************
JAVA_OPTS="{{ vitam_struct.jvm_opts.gc | default(gc_opts) }} {{ vitam_struct.jvm_opts.memory | default(memory_opts) }} {{ vitam_struct.jvm_opts.java | default(java_opts) }} -Dorg.owasp.esapi.resources={{ vitam_folder_conf }} -Dlogback.configurationFile={{ vitam_folder_conf }}/logback.xml -Dvitam.config.folder={{ vitam_folder_conf }} -Dvitam.data.folder={{ vitam_folder_data }} -Dvitam.tmp.folder={{ vitam_folder_tmp }} -Dvitam.log.folder={{ vitam_folder_log }} -Djava.security.properties={{ vitam_folder_conf }}/java.security -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath={{ vitam_folder_log }}{% if jvm_log %} -XX:+UnlockDiagnosticVMOptions -XX:+LogVMOutput -XX:LogFile={{ vitam_folder_log }}/jvm.log{% endif %} -XX:+UseG1GC {{ vitam_struct.jmx_exporter | default(jmx_exporter_opts) }}"
JAVA_ARGS="{{ vitam_folder_conf }}/{{ vitam_struct.vitam_component }}.conf"

8.2.1.2. Fichier /vitam/conf/<composant>/logback-access.xml

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
<?xml version="1.0" encoding="UTF-8"?>
<configuration>
{% if logback_rolling_policy|lower == "true" %}

  <appender name="FILE" class="ch.qos.logback.core.rolling.RollingFileAppender">
    <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
      <fileNamePattern>{{ vitam_folder_log }}/accesslog-{{ vitam_struct.vitam_component }}.%d{yyyy-MM-dd}.log</fileNamePattern>
      <maxHistory>{{ vitam_struct.access_retention_days | default(access_retention_days) }}</maxHistory>
          <totalSizeCap>{{ vitam_struct.access_total_size_cap | default(access_total_size_cap) }}</totalSizeCap>
    </rollingPolicy>
{% else %}}

  <appender name="FILE" class="ch.qos.logback.core.FileAppender">
    <file>{{ vitam_folder_log }}/accesslog-{{ vitam_struct.vitam_component }}.log</file>
    <append>true</append>
{% endif %}}
    <encoder>
      <pattern>%h %l %u %t "%r" %s %b "%i{Referer}" "%i{User-agent}" %D %i{X-Request-Id} %i{X-Tenant-Id} %i{X-Application-Id}</pattern>
    </encoder>
  </appender>
  <appender-ref ref="FILE" />
</configuration>

8.2.1.3. Fichier /vitam/conf/<composant>/logback.xml

  1
  2
  3
  4
  5
  6
  7
  8
  9
 10
 11
 12
 13
 14
 15
 16
 17
 18
 19
 20
 21
 22
 23
 24
 25
 26
 27
 28
 29
 30
 31
 32
 33
 34
 35
 36
 37
 38
 39
 40
 41
 42
 43
 44
 45
 46
 47
 48
 49
 50
 51
 52
 53
 54
 55
 56
 57
 58
 59
 60
 61
 62
 63
 64
 65
 66
 67
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
#jinja2: lstrip_blocks: True
<?xml version="1.0" encoding="UTF-8"?>
<configuration>

  <!-- Send debug messages to System.out -->
  <appender name="STDOUT" class="ch.qos.logback.core.ConsoleAppender">
    <!-- By default, encoders are assigned the type ch.qos.logback.classic.encoder.PatternLayoutEncoder -->
    <encoder>
      <pattern>%d{ISO8601} [[%thread]] [%X{X-Request-Id}] [%X{X-Tenant-Id}] %-5level %logger - %replace(%caller{1..2}){'Caller\+1   at |\n',''} : %msg %rootException%n</pattern>
    </encoder>
  </appender>

{% if logback_rolling_policy|lower == "true" %}
  <appender name="FILE" class="ch.qos.logback.core.rolling.RollingFileAppender">
    <rollingPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedRollingPolicy">
      <fileNamePattern>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
      <maxFileSize>{{ vitam_struct.logback_max_file_size | default(logback_max_file_size) }}</maxFileSize>
      <maxHistory>{{ vitam_struct.logback_total_size_cap.file.history_days | default(logback_total_size_cap.file.history_days) }}</maxHistory>
      <totalSizeCap>{{ vitam_struct.logback_total_size_cap.file.totalsize | default(logback_total_size_cap.file.totalsize) }}</totalSizeCap>
    </rollingPolicy>
{% else %}
  <appender name="FILE" class="ch.qos.logback.core.FileAppender">
    <file>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}.log</file>
    <append>true</append>
{% endif %}
    <encoder>
      <pattern>%d{ISO8601} [[%thread]] [%X{X-Request-Id}] [%X{X-Tenant-Id}] %-5level %logger - %replace(%caller{1..2}){'Caller\+1     at |\n',''} : %msg %rootException%n</pattern>
    </encoder>
  </appender>

{% if logback_rolling_policy|lower == "true" %}
  <appender name="SECURITY" class="ch.qos.logback.core.rolling.RollingFileAppender">
    <rollingPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedRollingPolicy">
      <fileNamePattern>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_security.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
      <maxFileSize>{{ vitam_struct.logback_max_file_size | default(logback_max_file_size) }}</maxFileSize>
      <maxHistory>{{ vitam_struct.logback_total_size_cap.security.history_days | default(logback_total_size_cap.security.history_days) }}</maxHistory>
      <totalSizeCap>{{ vitam_struct.logback_total_size_cap.security.totalsize | default(logback_total_size_cap.security.totalsize) }}</totalSizeCap>
    </rollingPolicy>
{% else %}
  <appender name="SECURITY" class="ch.qos.logback.core.FileAppender">
    <file>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_security.log</file>
    <append>true</append>
{% endif %}
    <encoder>
      <pattern>%d{ISO8601} [[%thread]] [%X{X-Request-Id}] [%X{X-Tenant-Id}] %-5level %logger - %replace(%caller{1..2}){'Caller\+1     at |\n',''} : %msg %rootException%n
      </pattern>
    </encoder>
  </appender>

{% if vitam_struct.vitam_component == 'storage' %}
  {% if logback_rolling_policy|lower == "true" %}
  <appender name="OFFERSYNC" class="ch.qos.logback.core.rolling.RollingFileAppender">
    <rollingPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedRollingPolicy">
      <fileNamePattern>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_offer_sync.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
      <maxFileSize>{{ vitam_struct.logback_max_file_size | default(logback_max_file_size) }}</maxFileSize>
      <maxHistory>{{ vitam_struct.logback_total_size_cap.offersync.history_days }}</maxHistory>
      <totalSizeCap>{{ vitam_struct.logback_total_size_cap.offersync.totalsize }}</totalSizeCap>
    </rollingPolicy>
  {% else %}
  <appender name="OFFERSYNC" class="ch.qos.logback.core.FileAppender">
    <file>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_offer_sync.log</file>
    <append>true</append>
  {% endif %}
    <encoder>
      <pattern>%d{ISO8601} [[%thread]] [%X{X-Request-Id}] [%X{X-Tenant-Id}] %-5level %logger - %replace(%caller{1..2}){'Caller\+1     at |\n',''} : %msg %rootException%n
      </pattern>
    </encoder>
  </appender>

  {% if logback_rolling_policy|lower == "true" %}
  <appender name="OFFERDIFF" class="ch.qos.logback.core.rolling.RollingFileAppender">
    <rollingPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedRollingPolicy">
      <fileNamePattern>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_offer_diff.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
      <maxFileSize>{{ vitam_struct.logback_max_file_size | default(logback_max_file_size) }}</maxFileSize>
      <maxHistory>{{ vitam_struct.logback_total_size_cap.offerdiff.history_days }}</maxHistory>
      <totalSizeCap>{{ vitam_struct.logback_total_size_cap.offerdiff.totalsize }}</totalSizeCap>
    </rollingPolicy>
  {% else %}
  <appender name="OFFERDIFF" class="ch.qos.logback.core.FileAppender">
    <file>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_offer_diff.log</file>
    <append>true</append>
  {% endif %}
    <encoder>
      <pattern>%d{ISO8601} [[%thread]] [%X{X-Request-Id}] [%X{X-Tenant-Id}] %-5level %logger - %replace(%caller{1..2}){'Caller\+1     at |\n',''} : %msg %rootException%n
      </pattern>
    </encoder>
  </appender>
{% endif %}

{% if vitam_struct.vitam_component == 'offer' and vitam_offers[offer_conf]["provider"] == 'tape-library' %}
  {% if logback_rolling_policy|lower == "true" %}
  <appender name="OFFER_TAPE" class="ch.qos.logback.core.rolling.RollingFileAppender">
    <rollingPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedRollingPolicy">
      <fileNamePattern>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_offer_tape.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
      <maxFileSize>{{ vitam_struct.logback_max_file_size | default(logback_max_file_size) }}</maxFileSize>
      <maxHistory>{{ vitam_struct.logback_total_size_cap.offer_tape.history_days }}</maxHistory>
      <totalSizeCap>{{ vitam_struct.logback_total_size_cap.offer_tape.totalsize }}</totalSizeCap>
    </rollingPolicy>
  {% else %}
  <appender name="OFFER_TAPE" class="ch.qos.logback.core.FileAppender">
    <file>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_offer_tape.log</file>
    <append>true</append>
  {% endif %}
    <encoder>
      <pattern>%d{ISO8601} [[%thread]] [%X{X-Request-Id}] [%X{X-Tenant-Id}] %-5level %logger - %replace(%caller{1..2}){'Caller\+1     at |\n',''} : %msg %rootException%n
      </pattern>
    </encoder>
  </appender>

  {% if logback_rolling_policy|lower == "true" %}
  <appender name="OFFER_TAPE_BACKUP" class="ch.qos.logback.core.rolling.RollingFileAppender">
    <rollingPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedRollingPolicy">
      <fileNamePattern>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_tape_backup.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
      <maxFileSize>{{ vitam_struct.logback_max_file_size | default(logback_max_file_size) }}</maxFileSize>
      <maxHistory>{{ vitam_struct.logback_total_size_cap.offer_tape_backup.history_days }}</maxHistory>
      <totalSizeCap>{{ vitam_struct.logback_total_size_cap.offer_tape_backup.totalsize }}</totalSizeCap>
    </rollingPolicy>
  {% else %}
  <appender name="OFFER_TAPE_BACKUP" class="ch.qos.logback.core.FileAppender">
    <file>{{ vitam_folder_log }}/{{ vitam_struct.vitam_component }}_tape_backup.log</file>
    <append>true</append>
  {% endif %}
    <encoder>
      <pattern>%d{ISO8601} [[%thread]] [%X{X-Request-Id}] [%X{X-Tenant-Id}] %-5level %logger - %replace(%caller{1..2}){'Caller\+1     at |\n',''} : %msg %rootException%n
      </pattern>
    </encoder>
  </appender>
{% endif %}

  <appender name="SYSLOG" class="ch.qos.logback.classic.net.SyslogAppender">
    <syslogHost>localhost</syslogHost>
    <facility>{{ vitam_defaults.syslog_facility | default('local0') }}</facility>
    <suffixPattern>vitam-{{ vitam_struct.vitam_component }}: %d{ISO8601} [[%thread]] [%X{X-Request-Id}] [%X{X-Tenant-Id}] %-5level %logger - %replace(%caller{1..2}){'Caller\+1     at |\n',''} : %msg %rootException%n</suffixPattern>
  </appender>
  <!-- By default, the level of the root level is set to TRACE -->
  <root level="{{ vitam_struct.log_level | default(vitam_defaults.services.log_level) }}">
    <!-- <appender-ref ref="STDOUT" /> -->
    <appender-ref ref="FILE" />
    <appender-ref ref="SYSLOG" />
  </root>

  <logger name="org.eclipse.jetty" level="WARN"/>
  <logger name="fr.gouv.vitam.storage.engine.server.logbook.StorageLogbookMock" level="INFO"/>
  <logger name="fr.gouv.vitam.metadata.core.graph.StoreGraphService" level="INFO"/>
  <logger name="fr.gouv.vitam.metadata.core.graph.GraphComputeServiceImpl" level="INFO"/>
  <logger name="fr.gouv.vitam.common" level="WARN" />
{% if vitam_defaults.reconstruction.log_level is defined or reconstruction.log_level is defined %}
  <logger name="fr.gouv.vitam.metadata.core.reconstruction.ReconstructionService" level="{{ vitam_struct.reconstruction.log_level | default(reconstruction.log_level) }}"/>
  <logger name="fr.gouv.vitam.metadata.core.reconstruction.RestoreBackupService" level="{{ vitam_struct.reconstruction.log_level | default(reconstruction.log_level) }}"/>
  <logger name="fr.gouv.vitam.logbook.common.server.reconstruction.ReconstructionService" level="{{ vitam_struct.reconstruction.log_level | default(reconstruction.log_level) }}"/>
  <logger name="fr.gouv.vitam.logbook.common.server.reconstruction.RestoreBackupService" level="{{ vitam_struct.reconstruction.log_level | default(reconstruction.log_level) }}"/>
  <logger name="fr.gouv.vitam.functional.administration.common.impl.ReconstructionServiceImpl" level="{{ vitam_struct.reconstruction.log_level | default(reconstruction.log_level) }}"/>
  <logger name="fr.gouv.vitam.functional.administration.common.impl.RestoreBackupServiceImpl" level="{{ vitam_struct.reconstruction.log_level | default(reconstruction.log_level) }}"/>
{% endif %}

{% if performance_logger|lower == "true" %}
  <logger name="fr.gouv.vitam.common.performance.PerformanceLogger" level="DEBUG" additivity="false" >
    <appender-ref ref="SYSLOG" />
  </logger>
{% endif %}

  <logger name="fr.gouv.vitam.common.alert.AlertServiceImpl" level="INFO">
    <appender-ref ref="SECURITY" />
  </logger>

{% if vitam_struct.vitam_component == 'storage' %}
  <logger name="fr.gouv.vitam.storage.engine.server.offersynchronization" level="INFO">
    <appender-ref ref="OFFERSYNC" />
  </logger>
  <logger name="fr.gouv.vitam.storage.engine.server.offerdiff" level="INFO">
    <appender-ref ref="OFFERDIFF" />
  </logger>
{% endif %}

{% if vitam_struct.vitam_component == 'offer' %}
  <logger name="fr.gouv.vitam.storage.offers.tape.process.ProcessExecutor" level="INFO" additivity="false" >
    <appender-ref ref="OFFER_TAPE" />
  </logger>

  <logger name="fr.gouv.vitam.storage.offers.tape.utils.BackupLogInformation" level="INFO" additivity="false" >
    <appender-ref ref="OFFER_TAPE_BACKUP" />
  </logger>
{% endif %}

{% if vitam_struct.vitam_component == 'metadata' %}
  <logger name="fr.gouv.vitam.metadata.core.migration" level="INFO"/>
{% endif %}

{% if vitam_struct.vitam_component == 'scheduler' %}
  <logger name="org.quartz" level="INFO"/>
  <logger name="fr.gouv.vitam.scheduler" level="INFO"/>
{% endif %}

</configuration>

8.2.1.4. Fichier /vitam/conf/<composant>/jetty-config.xml

  1
  2
  3
  4
  5
  6
  7
  8
  9
 10
 11
 12
 13
 14
 15
 16
 17
 18
 19
 20
 21
 22
 23
 24
 25
 26
 27
 28
 29
 30
 31
 32
 33
 34
 35
 36
 37
 38
 39
 40
 41
 42
 43
 44
 45
 46
 47
 48
 49
 50
 51
 52
 53
 54
 55
 56
 57
 58
 59
 60
 61
 62
 63
 64
 65
 66
 67
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
<?xml version="1.0"?>
<!DOCTYPE Configure PUBLIC "-//Jetty//Configure//EN" "http://www.eclipse.org/jetty/configure_9_0.dtd">

<!-- =============================================================== -->
<!-- Documentation of this file format can be found at:              -->
<!-- http://wiki.eclipse.org/Jetty/Reference/jetty.xml_syntax        -->
<!--                                                                 -->
<!-- Additional configuration files are available in $JETTY_HOME/etc -->
<!-- and can be mixed in. See start.ini file for the default         -->
<!-- configuration files.                                            -->
<!--                                                                 -->
<!-- For a description of the configuration mechanism, see the       -->
<!-- output of:                                                      -->
<!--   java -jar start.jar -?                                        -->
<!-- =============================================================== -->

<!-- =============================================================== -->
<!-- Configure a Jetty Server instance with an ID "Server"           -->
<!-- Other configuration files may also configure the "Server"       -->
<!-- ID, in which case they are adding configuration to the same     -->
<!-- instance.  If other configuration have a different ID, they     -->
<!-- will create and configure another instance of Jetty.            -->
<!-- Consult the javadoc of o.e.j.server.Server for all              -->
<!-- configuration that may be set here.                             -->
<!-- =============================================================== -->
<Configure id="Server" class="org.eclipse.jetty.server.Server">


    <!-- =========================================================== -->
    <!-- Add shared Scheduler instance                               -->
    <!-- =========================================================== -->
    <Call name="addBean">
        <Arg>
            <New class="org.eclipse.jetty.util.thread.ScheduledExecutorScheduler"/>
        </Arg>
    </Call>

    <!-- =========================================================== -->
    <!-- Http Configuration.                                         -->
    <!-- This is a common configuration instance used by all         -->
    <!-- connectors that can carry HTTP semantics (HTTP, HTTPS, SPDY)-->
    <!-- It configures the non wire protocol aspects of the HTTP     -->
    <!-- semantic.                                                   -->
    <!--                                                             -->
    <!-- This configuration is only defined here and is used by      -->
    <!-- reference from the jetty-http.xml, jetty-https.xml and      -->
    <!-- jetty-spdy.xml configuration files which instantiate the    -->
    <!-- connectors.                                                 -->
    <!--                                                             -->
    <!-- Consult the javadoc of o.e.j.server.HttpConfiguration       -->
    <!-- for all configuration that may be set here.                 -->
    <!-- =========================================================== -->
    <New id="httpConfig" class="org.eclipse.jetty.server.HttpConfiguration">
        <Set name="secureScheme">http</Set>
        <Set name="securePort">8443</Set>
        <Set name="outputBufferSize">32768</Set>
        <Set name="requestHeaderSize">8192</Set>
        <Set name="responseHeaderSize">8192</Set>
        <Set name="sendServerVersion">false</Set>
        <Set name="sendDateHeader">false</Set>
        <Set name="headerCacheSize">512</Set>

        <!-- Uncomment to enable handling of X-Forwarded- style headers -->
        <!-- <Call name="addCustomizer">
          <Arg><New class="org.eclipse.jetty.server.ForwardedRequestCustomizer"/></Arg>
        </Call> -->

    </New>

    <!-- ======================= Original Connector ======================= -->
    <!-- <Call name="addConnector">
        <Arg>
            <New class="org.eclipse.jetty.server.ServerConnector">
                <Arg name="server"><Ref refid="Server" /></Arg>
                <Arg name="factories">
                    <Array type="org.eclipse.jetty.server.ConnectionFactory">
                        <Item>
                            <New class="org.eclipse.jetty.server.HttpConnectionFactory">
                                <Arg name="config"><Ref refid="httpConfig" /></Arg>
                            </New>
                        </Item>
                    </Array>
                </Arg>
                <Set name="port">{{ vitam_struct.port_service }}</Set>
                <Set name="idleTimeout">
                    <Property name="http.timeout" default="{{ vitam_defaults.services.port_service_timeout }}"/>
                </Set>
            </New>
        </Arg>
    </Call> -->

    <!-- =========================================================== -->
    <!-- Set the default handler structure for the Server            -->
    <!-- A handler collection is used to pass received requests to   -->
    <!-- both the ContextHandlerCollection, which selects the next   -->
    <!-- handler by context path and virtual host, and the           -->
    <!-- DefaultHandler, which handles any requests not handled by   -->
    <!-- the context handlers.                                       -->
    <!-- Other handlers may be added to the "Handlers" collection,   -->
    <!-- for example the jetty-requestlog.xml file adds the          -->
    <!-- RequestLogHandler after the default handler                 -->
    <!-- =========================================================== -->
    <Set name="handler">
        <New id="Handlers" class="org.eclipse.jetty.server.handler.HandlerCollection">
            <Set name="handlers">
                <Array type="org.eclipse.jetty.server.Handler">
                    <Item>
                        <New id="Contexts" class="org.eclipse.jetty.server.handler.ContextHandlerCollection"/>
                    </Item>
                    <Item>
                        <New id="DefaultHandler" class="org.eclipse.jetty.server.handler.DefaultHandler"/>
                    </Item>
                </Array>
            </Set>
        </New>
    </Set>

    <Set name="RequestLog">
            <New id="RequestLogImpl" class="ch.qos.logback.access.jetty.VitamRequestLogImpl">
                <Set name="fileName">{{ vitam_folder_conf }}/logback-access.xml</Set>
            </New>
    </Set>
    <Ref refid="RequestLogImpl">
              <Call name="start"/>
    </Ref>

    <!-- =========================================================== -->
    <!-- extra server options                                        -->
    <!-- =========================================================== -->
    <Set name="stopAtShutdown">true</Set>
    <Set name="stopTimeout">5000</Set>
    <Set name="dumpAfterStart">false</Set>
    <Set name="dumpBeforeStop">false</Set>

{% if vitam_struct.https_enabled==true %}
   <New id="httpsConfig" class="org.eclipse.jetty.server.HttpConfiguration">
        <Set name="sendServerVersion">false</Set>
        <Set name="sendDateHeader">false</Set>
        <Call name="addCustomizer">
               <Arg>
                   <New class="org.eclipse.jetty.server.SecureRequestCustomizer">
                       <Arg name="sniHostCheck" type="boolean">
                           <Property name="jetty.ssl.sniHostCheck" default="false" />
                       </Arg>
                   </New>
               </Arg>
        </Call>
    </New>
        <New id="sslContextFactory" class="org.eclipse.jetty.util.ssl.SslContextFactory$Server">
        <Set name="KeyStorePath">{{ vitam_folder_conf }}/keystore_{{ vitam_struct.vitam_component }}.jks</Set>
        <Set name="KeyStorePassword">{{ password_keystore }}</Set>
        <Set name="KeyManagerPassword">{{ password_manager_keystore }}</Set>
        <Set name="TrustStorePath">{{ vitam_folder_conf }}/truststore_{{ vitam_struct.vitam_component }}.jks</Set>
        <Set name="TrustStorePassword">{{ password_truststore }}</Set>
        <Set name="TrustStoreType">JKS</Set>
        <Set name="NeedClientAuth">false</Set>
        <Set name="WantClientAuth">true</Set>
      <Set name="IncludeCipherSuites">
        <Array type="String">
          <Item>TLS_ECDHE.*</Item>
          <Item>TLS_DHE_RSA.*</Item>
        </Array>
      </Set>
       <Set name="IncludeProtocols">
        <Array type="String">
          <Item>TLSv1.2</Item>
        </Array>
      </Set>
      <Set name="ExcludeCipherSuites">
        <Array type="String">
          <Item>.*NULL.*</Item>
          <Item>.*RC4.*</Item>
          <Item>.*MD5.*</Item>
          <Item>.*DES.*</Item>
          <Item>.*DSS.</Item>
          <Item>TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA</Item>
          <Item>TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA</Item>
          <Item>TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA</Item>
          <Item>TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA</Item>
          <Item>TLS_DHE_RSA_WITH_AES_256_CBC_SHA</Item>
          <Item>TLS_DHE_RSA_WITH_AES_128_CBC_SHA</Item>
        </Array>
      </Set>
        <Set name="UseCipherSuitesOrder">true</Set>
        <Set name="RenegotiationAllowed">true</Set>
    </New>
    <New id="sslConnectionFactory" class="org.eclipse.jetty.server.SslConnectionFactory">
        <Arg name="sslContextFactory">
            <Ref refid="sslContextFactory" />
        </Arg>
        <Arg name="next">http/1.1</Arg>
    </New>
    <New id="businessConnector" class="org.eclipse.jetty.server.ServerConnector">
        <Arg name="server">
            <Ref refid="Server" />
        </Arg>
        <Arg name="factories">
            <Array type="org.eclipse.jetty.server.ConnectionFactory">
                <Item>
                    <Ref refid="sslConnectionFactory" />
                </Item>
                <Item>
                    <New class="org.eclipse.jetty.server.HttpConnectionFactory">
                        <Arg name="config">
                            <Ref refid="httpsConfig" />
                        </Arg>
                    </New>
                </Item>
            </Array>
        </Arg>
        <Set name="host">{{ ip_service }}</Set>
        <Set name="port">
            <SystemProperty name="jetty.port" default="{{ vitam_struct.port_service }}"/>
        </Set>
        <Set name="name">business</Set>
    </New>

{% else %}

    <!-- =========================================================== -->
    <!-- Connector for API business                                  -->
    <!-- Attach all ContextHanlder except Admin                      -->
    <!-- =========================================================== -->

    <New id="businessConnector" class="org.eclipse.jetty.server.ServerConnector">
        <Arg name="server"><Ref refid="Server" /></Arg>
        <Arg name="factories">
            <Array type="org.eclipse.jetty.server.ConnectionFactory">
                <Item>
                    <New class="org.eclipse.jetty.server.HttpConnectionFactory">
                        <Arg name="config"><Ref refid="httpConfig" /></Arg>
                    </New>
                </Item>
            </Array>
        </Arg>
        <Set name="host">{{ ip_service }}</Set>
        <Set name="port">{{ vitam_struct.port_service }}</Set>
        <Set name="name">business</Set>
        <Set name="idleTimeout">
            <Property name="http.timeout" default="{{ vitam_defaults.services.port_service_timeout }}"/>
        </Set>
    </New>

{% endif %}

    <!-- =========================================================== -->
    <!-- Connector for API Admin                                     -->
    <!-- Attach all ContextHanlder                                   -->
    <!-- =========================================================== -->

    <New id="adminConnector" class="org.eclipse.jetty.server.ServerConnector">
        <Arg name="server"><Ref refid="Server" /></Arg>
        <Arg name="factories">
            <Array type="org.eclipse.jetty.server.ConnectionFactory">
                <Item>
                    <New class="org.eclipse.jetty.server.HttpConnectionFactory">
                        <Arg name="config"><Ref refid="httpConfig" /></Arg>
                    </New>
                </Item>
            </Array>
        </Arg>
        <Set name="host">{{ ip_admin }}</Set>
        <Set name="port">{{ vitam_struct.port_admin }}</Set>
        <Set name="name">admin</Set>
        <Set name="idleTimeout">
            <Property name="http.timeout" default="{{ vitam_defaults.services.port_service_timeout }}"/>
        </Set>
    </New>

    <Call name="setConnectors">
        <Arg>
             <Array type="org.eclipse.jetty.server.ServerConnector">
                <Item>
                    <Ref refid="businessConnector" />
                </Item>
                <Item>
                    <Ref refid="adminConnector" />
                </Item>
            </Array>
        </Arg>
    </Call>

</Configure>

8.2.1.5. Fichier /vitam/conf/<composant>/logbook-client.conf

Ce fichier permet de configurer l’appel au composant logbook.

8.2.1.6. Fichier /vitam/conf/<composant>/server-identity.conf

1
2
3
identityName: {{ ansible_nodename }}
identityRole: {{ vitam_struct.vitam_component }}
identitySiteId: {{ vitam_site_id }}

8.2.1.7. Fichier /vitam/conf/<composant>/antisamy-esapi.xml

8.2.1.8. Fichier /vitam/conf/<composant>/vitam.conf

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
secret : {{ plateforme_secret }}
filterActivation : {{ vitam_struct.secret_platform }}
{% if vitam_struct.vitam_component == vitam.processing.vitam_component %}
distributeurBatchSize: 800
workerBulkSize: 16
{% endif %}
{% if vitam_struct.vitam_component == vitam.metadata.vitam_component %}
storeGraphElementsPerFile: 10000
storeGraphOverlapDelay: 300
expireCacheEntriesDelay: 300
deleteIncompleteReconstructedUnitDelay: 2592000
migrationBulkSize: 10000
workspaceFreespaceThreshold: {{ vitam.metadata.workspaceFreespaceThreshold | default(25) }}
{% endif %}
distributionThreshold : 100000
eliminationAnalysisThreshold : 100000
eliminationActionThreshold : 10000
computedInheritedRulesThreshold : 100000000
intervalDelayCheckIdle : 5000
maxDelayUnusedConnection : 5000
delayValidationAfterInactivity : 2500
tenants: [ "{{ vitam_tenant_ids | join('", "') }}" ]
adminTenant : {{ vitam_tenant_admin | default(1) }}
forceChunkModeInputStream : {{ vitam_defaults.vitam_force_chunk_mode | default(false) }}

{% if vitam_struct.vitam_component == vitam.worker.vitam_component %}
reclassificationMaxBulkThreshold: 1000
reclassificationMaxUnitsThreshold: 10000
reclassificationMaxGuildListSizeInLogbookOperation: 1000
queriesThreshold: {{ vitam.worker.queriesThreshold | default(100000) }}
bulkAtomicUpdateBatchSize: {{ vitam.worker.bulkAtomicUpdateBatchSize | default(100) }}
bulkAtomicUpdateThreadPoolSize: {{ vitam.worker.bulkAtomicUpdateThreadPoolSize | default(8) }}
bulkAtomicUpdateThreadPoolQueueSize: {{ vitam.worker.bulkAtomicUpdateThreadPoolQueueSize | default(16) }}

binarySizePlatformThreshold: # 1 Go
 limit: {{ vitam.worker.binarySizePlatformThreshold | default(1) }}
 sizeUnit: {{ vitam.worker.binarySizePlatformThresholdSizeUnit | default('GIGABYTE') }}

binarySizeTenantThreshold: # exemple for tenant 0 max dip/transfer size is 20 Mo, true means can exceed tenant threshold.
#  - tenant: 0
#    limit: 20
#    sizeUnit: MEGABYTE
#    authorize: true

{% endif %}

keywordMaxLength: 32766
textMaxLength: 32766

classificationLevel :
  allowList : [{% for classification in classificationList %}{{ classification }}{% if not loop.last %},{% endif %}{% endfor %}]
  authorizeNotDefined: {{ classificationLevelOptional }}

indexInheritedRulesWithAPIV2OutputByTenant: [ "{{ vitam.worker.api_output_index_tenants | join('", "') }}" ]
indexInheritedRulesWithRulesIdByTenant: [ "{{ vitam.worker.rules_index_tenants | join('", "') }}" ]

environmentName: {{ vitam_prefix_offer|default(vitam_site_name) }}

acceptableRequestTime: {{ acceptableRequestTime|default(10) }}
criticalRequestTime: {{ criticalRequestTime|default(60) }}
requestTimeAlertThrottlingDelay: {{ requestTimeAlertThrottlingDelay|default(60) }}

# Ontology cache settings (max entries in cache & retention timeout in seconds)
ontologyCacheMaxEntries: {{ vitam_defaults.ontologyCacheMaxEntries | default(100) }}
ontologyCacheTimeoutInSeconds: {{ vitam_defaults.ontologyCacheTimeoutInSeconds | default(300) }}

# Elasticsearch scroll timeout settings
elasticSearchScrollTimeoutInMilliseconds: {{ vitam_defaults.elasticSearchScrollTimeoutInMilliseconds | default(300000) }}

{% if vitam_struct.vitam_component == vitam.processing.vitam_component %}
processEngineWaitForStepTimeout: 172800
{% endif %}

Ce fichier permet de définir les variables d’environnement vitam.

  • binarySizePlatformThreshold est le seuil de plate-forme du poids binaire max autorisé pour un DIP.
    elle comporte deux clés :

    limit : le seuil sizeUnit : l’unité de taille (GIGABYTE / MEGABYTE / KILOBYTE / BYTE)

    par défaut le seuil est 1 Go. exemple :

    binarySizePlatformThreshold:
        limit: 1
        sizeUnit: GIGABYTE / MEGABYTE / KILOBYTE / BYTE
    
  • binarySizeTenantThreshold est une liste qui constitue l’ensemble des seuils du poids binaire max autorisé pour un DIP par tenant.
    Cette liste comporte 4 clés :

    tenant : le tenant limit : le seuil sizeUnit : l’unité de taille (GIGABYTE / MEGABYTE / KILOBYTE / BYTE) authorise : true si l’utilisateur peut excéder le seuil prédéfini.

    exemple :

    binarySizeTenantThreshold:
       - tenant: 0
          limit: 20
          sizeUnit: MEGABYTE
          authorize: false
       - tenant: 1
          limit: 100
          sizeUnit: MEGABYTE
          authorize: true
    

8.2.1.9. Fichier /vitam/conf/<composant>/java.security

1
2
3
4
5
# Use Bouncy Castle Provider when it is available
security.provider.9=org.bouncycastle.jce.provider.BouncyCastleProvider

# Override the default list of Centos 7 that disable Elliptic Curved Based Algorithms
jdk.tls.disabledAlgorithms="SSLv3, TLSv1, TLSv1.1, RC4, MD5withRSA, DH keySize < 768,RSA keySize < 2048"