Transform the "application.properties" file to "application.yml" and optimize the property-trees.

This commit is contained in:
Lampros Smyrnaios 2023-03-20 15:23:00 +02:00
parent 17a6c120dd
commit f835a752bf
3 changed files with 75 additions and 76 deletions

View File

@ -79,7 +79,7 @@ public class ParquetFileUtils {
public ParquetFileUtils(@Value("${hdfs.baseUrl}") String webHDFSBaseUrl,
@Value("${hdfs.httpAuth}") String hdfsHttpAuthString, @Value("${hdfs.userName}") String hdfsUserName, @Value("${hdfs.password}") String hdfsPassword, @Value("${output.parquetLocalDirectoryPath}") String parquetBaseDirectoryPath,
@Value("${hdfs.httpAuth}") String hdfsHttpAuthString, @Value("${hdfs.userName}") String hdfsUserName, @Value("${hdfs.password}") String hdfsPassword, @Value("${services.pdfaggregation.controller.parquetLocalDirectoryPath}") String parquetBaseDirectoryPath,
@Value("${hdfs.parquetRemoteBaseDirectoryPath}") String hdfsParquetBaseDir,
@Value("${services.pdfaggregation.controller.isTestEnvironment}") boolean isTestEnvironment, FileUtils fileUtils) throws IOException
{

View File

@ -1,75 +0,0 @@
# HTTP CONFIGURATION
server.port = 1880
# Server api path
server.servlet.context-path=/api
# Service config
services.pdfaggregation.controller.isTestEnvironment = false
# In case the "isTestEnvironment" is "true", the "testDatabase" below and all its tables are created (if not exist).
# The tables "datasource", "publication", "publication_pids" and "publication_urls" are filled with the data from the same tables existing in the "initialDatabase", if they don't exist.
# In case the "isTestEnvironment" is "false", the "initialDatabase" is used. The Controller assumes that the above 4 tables are present, and only creates, if they don't exist, the following tables:
# "assignment", "attempt" and "payload", which are populated during execution.
services.pdfaggregation.controller.db.initialDatabaseName = pdfaggregation_i
services.pdfaggregation.controller.db.testDatabaseName = pdfaggregationdatabase_new_s3_names
services.pdfaggregation.controller.baseFilesLocation = tmp/
services.pdfaggregation.controller.maxAttemptsPerRecord = 3
services.pdfaggregation.controller.assignmentLimit = 10000
services.pdfaggregation.controller.s3.endpoint = xa
services.pdfaggregation.controller.s3.accessKey = xa
services.pdfaggregation.controller.s3.secretKey = xa
services.pdfaggregation.controller.s3.region = xa
services.pdfaggregation.controller.s3.bucketName = xa
services.pdfaggregation.controller.s3.shouldEmptyBucket = false
services.pdfaggregation.controller.s3.shouldShowAllS3Buckets = true
# Database
spring.datasource.url=jdbc:impala://iis-cdh5-test-gw.ocean.icm.edu.pl:21050/
spring.datasource.username=
spring.datasource.password=
spring.datasource.driver-class-name=com.cloudera.impala.jdbc41.Driver
spring.datasource.hikari.pool-name=ControllerPool
spring.datasource.hikari.maximumPoolSize=20
spring.datasource.hikari.maxLifetime=1800000
spring.datasource.hikari.minimumIdle=4
spring.datasource.hikari.connectionTimeout=30000
spring.datasource.hikari.idleTimeout=600000
# LOGGING LEVELS
logging.level.root=INFO
logging.level.org.springframework.web=INFO
logging.level.org.springframework.security=WARN
logging.level.org.apache.hadoop.io.compress=WARN
logging.level.eu.openaire.urls_controller=DEBUG
spring.output.ansi.enabled=always
# Parquet settings
hdfs.baseUrl=https://iis-cdh5-test-gw.ocean.icm.edu.pl/webhdfs/v1
# HTTP-Authorization --> Authorization: Basic Base64Encode(username:password)
# Give the credentials by either giving the Http-Auth-string AND the username (used as parameter in the WebHdfs-requests)
# Or by giving the username AND the password, in order for the program to crete the auth-String programmatically.
# The first approach is intended for more privacy, while the second for more ease. Either way, all three should be uncommented, no matter which ones are used.
hdfs.httpAuth=
hdfs.userName=
hdfs.password=
output.parquetLocalDirectoryPath=${services.pdfaggregation.controller.baseFilesLocation}parquetFiles/
hdfs.parquetRemoteBaseDirectoryPath=/tmp/parquet_uploads/
## MULTIPART (MultipartProperties)
# Enable multipart uploads
spring.servlet.multipart.enabled=true
# Threshold after which files are written to disk.
spring.servlet.multipart.file-size-threshold=2KB
# Max file size.
spring.servlet.multipart.max-file-size=200MB
# Max Request Size
spring.servlet.multipart.max-request-size=215MB

View File

@ -0,0 +1,74 @@
server:
port: 1880
servlet:
context-path: /api
services:
pdfaggregation:
controller:
isTestEnvironment: false
# In case the "isTestEnvironment" is "true", the "testDatabase" below and all its tables are created (if not exist).
# The tables "datasource", "publication", "publication_pids" and "publication_urls" are filled with the data from the same tables existing in the "initialDatabase", if they don't exist.
# In case the "isTestEnvironment" is "false", the "initialDatabase" is used. The Controller assumes that the above 4 tables are present, and only creates, if they don't exist, the following tables:
# "assignment", "attempt" and "payload", which are populated during execution.
db:
initialDatabaseName: pdfaggregation_i
testDatabaseName: pdfaggregationdatabase_new_s3_names
assignmentLimit: 10000
maxAttemptsPerRecord: 3
baseFilesLocation: tmp/
parquetLocalDirectoryPath: ${services.pdfaggregation.controller.baseFilesLocation}parquetFiles/
s3:
endpoint: XA
accessKey: XA
secretKey: XA
region: XA
bucketName: XA
shouldEmptyBucket: false
shouldShowAllS3Buckets: true
spring:
datasource:
driver-class-name: com.cloudera.impala.jdbc41.Driver
url: XA
username: ''
password: ''
hikari:
connectionTimeout: 30000
idleTimeout: 600000
maxLifetime: 1800000
maximumPoolSize: 20
minimumIdle: 4
pool-name: ControllerPool
output:
ansi:
enabled: always
hdfs:
baseUrl: XA
userName: XA
password: XA
httpAuth: ''
# HTTP-Authorization --> Authorization: Basic Base64Encode(username:password)
# Give the credentials by either giving the Http-Auth-string AND the username (used as parameter in the WebHdfs-requests)
# Or by giving the username AND the password, in order for the program to crete the auth-String programmatically.
# The first approach is intended for more privacy, while the second for more ease. Either way, all three should be uncommented, no matter which ones are used.
parquetRemoteBaseDirectoryPath: /tmp/parquet_uploads/
logging:
level:
root: INFO
eu:
openaire:
urls_controller: DEBUG
org:
springframework:
security: WARN
web: INFO
apache:
hadoop:
io:
compress: WARN