From 7ecbd2e269eb1ec971d2ef3ea2e4bd7ff953c2d2 Mon Sep 17 00:00:00 2001 From: Daniel Bell Date: Wed, 17 Jan 2024 09:02:42 +0100 Subject: [PATCH] Scalafmt --- .../kernel/utils/HandlebarsExpander.scala | 21 ++++++++++++------- .../storage/files/batch/BatchCopySuite.scala | 18 ++++++++++------ .../files/routes/FilesRoutesSpec.scala | 20 ++++++++++-------- 3 files changed, 36 insertions(+), 23 deletions(-) diff --git a/delta/kernel/src/main/scala/ch/epfl/bluebrain/nexus/delta/kernel/utils/HandlebarsExpander.scala b/delta/kernel/src/main/scala/ch/epfl/bluebrain/nexus/delta/kernel/utils/HandlebarsExpander.scala index 9342026658..f4738b131f 100644 --- a/delta/kernel/src/main/scala/ch/epfl/bluebrain/nexus/delta/kernel/utils/HandlebarsExpander.scala +++ b/delta/kernel/src/main/scala/ch/epfl/bluebrain/nexus/delta/kernel/utils/HandlebarsExpander.scala @@ -6,16 +6,21 @@ import scala.jdk.CollectionConverters._ class HandlebarsExpander { - private val handleBars = new Handlebars().`with`(EscapingStrategy.NOOP).registerHelper("empty", new Helper[Iterable[_]] { - override def apply(context: Iterable[_], options: Options): CharSequence = { - context.iterator.isEmpty match { - case true => options.fn() - case false => options.inverse() + private val handleBars = new Handlebars() + .`with`(EscapingStrategy.NOOP) + .registerHelper( + "empty", + new Helper[Iterable[_]] { + override def apply(context: Iterable[_], options: Options): CharSequence = { + context.iterator.isEmpty match { + case true => options.fn() + case false => options.inverse() + } + } } - } - }) + ) - def expand(templateText: String,attributes: Map[String, Any]) = { + def expand(templateText: String, attributes: Map[String, Any]) = { if (attributes.isEmpty) { templateText } else { diff --git a/delta/plugins/storage/src/test/scala/ch/epfl/bluebrain/nexus/delta/plugins/storage/files/batch/BatchCopySuite.scala b/delta/plugins/storage/src/test/scala/ch/epfl/bluebrain/nexus/delta/plugins/storage/files/batch/BatchCopySuite.scala index 61e7cf38f0..920d331320 100644 --- a/delta/plugins/storage/src/test/scala/ch/epfl/bluebrain/nexus/delta/plugins/storage/files/batch/BatchCopySuite.scala +++ b/delta/plugins/storage/src/test/scala/ch/epfl/bluebrain/nexus/delta/plugins/storage/files/batch/BatchCopySuite.scala @@ -45,7 +45,8 @@ class BatchCopySuite extends NexusSuite with StorageFixtures with Generators wit test("successfully perform disk copy") { val events = ListBuffer.empty[Event] - val (sourceFileRes, sourceStorage) = genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata)) + val (sourceFileRes, sourceStorage) = + genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata)) val (user, aclCheck) = userAuthorizedOnProjectStorage(sourceStorage.value) val batchCopy = mkBatchCopy( @@ -74,7 +75,8 @@ class BatchCopySuite extends NexusSuite with StorageFixtures with Generators wit test("successfully perform remote disk copy") { val events = ListBuffer.empty[Event] - val (sourceFileRes, sourceStorage) = genFileResourceAndStorage(sourceFileId, sourceProj.context, remoteVal, Some(stubbedFileMetadata)) + val (sourceFileRes, sourceStorage) = + genFileResourceAndStorage(sourceFileId, sourceProj.context, remoteVal, Some(stubbedFileMetadata)) val (user, aclCheck) = userAuthorizedOnProjectStorage(sourceStorage.value) val batchCopy = mkBatchCopy( @@ -110,7 +112,8 @@ class BatchCopySuite extends NexusSuite with StorageFixtures with Generators wit test("fail if a source storage is different to destination storage") { val events = ListBuffer.empty[Event] - val (sourceFileRes, sourceStorage) = genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata)) + val (sourceFileRes, sourceStorage) = + genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata)) val (user, aclCheck) = userAuthorizedOnProjectStorage(sourceStorage.value) val batchCopy = mkBatchCopy( @@ -129,7 +132,8 @@ class BatchCopySuite extends NexusSuite with StorageFixtures with Generators wit test("fail if user does not have read access on a source file's storage") { val events = ListBuffer.empty[Event] - val (sourceFileRes, sourceStorage) = genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata)) + val (sourceFileRes, sourceStorage) = + genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata)) val user = genUser() val aclCheck = AclSimpleCheck((user, AclAddress.fromProject(sourceProj.ref), Set())).accepted @@ -148,7 +152,8 @@ class BatchCopySuite extends NexusSuite with StorageFixtures with Generators wit test("fail if a single source file exceeds max size for destination storage") { val events = ListBuffer.empty[Event] - val (sourceFileRes, sourceStorage) = genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata), 1000L) + val (sourceFileRes, sourceStorage) = + genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata), 1000L) val (user, aclCheck) = userAuthorizedOnProjectStorage(sourceStorage.value) val batchCopy = mkBatchCopy( @@ -172,7 +177,8 @@ class BatchCopySuite extends NexusSuite with StorageFixtures with Generators wit val capacity = 10L val statEntry = StorageStatEntry(files = 10L, spaceUsed = 1L) val spaceLeft = capacity - statEntry.spaceUsed - val (sourceFileRes, sourceStorage) = genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata), fileSize) + val (sourceFileRes, sourceStorage) = + genFileResourceAndStorage(sourceFileId, sourceProj.context, diskVal, Some(stubbedFileMetadata), fileSize) val (user, aclCheck) = userAuthorizedOnProjectStorage(sourceStorage.value) val batchCopy = mkBatchCopy( diff --git a/delta/plugins/storage/src/test/scala/ch/epfl/bluebrain/nexus/delta/plugins/storage/files/routes/FilesRoutesSpec.scala b/delta/plugins/storage/src/test/scala/ch/epfl/bluebrain/nexus/delta/plugins/storage/files/routes/FilesRoutesSpec.scala index baf9356f52..f24751aef5 100644 --- a/delta/plugins/storage/src/test/scala/ch/epfl/bluebrain/nexus/delta/plugins/storage/files/routes/FilesRoutesSpec.scala +++ b/delta/plugins/storage/src/test/scala/ch/epfl/bluebrain/nexus/delta/plugins/storage/files/routes/FilesRoutesSpec.scala @@ -12,10 +12,10 @@ import ch.epfl.bluebrain.nexus.delta.kernel.http.MediaTypeDetectorConfig import ch.epfl.bluebrain.nexus.delta.plugins.elasticsearch.model.ResourcesSearchParams.FileUserMetadata import ch.epfl.bluebrain.nexus.delta.plugins.storage.files.model.Digest.ComputedDigest import ch.epfl.bluebrain.nexus.delta.plugins.storage.files.model.{FileAttributes, FileId, FileRejection} -import ch.epfl.bluebrain.nexus.delta.plugins.storage.files.{FileFixtures, Files, FilesConfig, permissions, contexts => fileContexts} +import ch.epfl.bluebrain.nexus.delta.plugins.storage.files.{contexts => fileContexts, permissions, FileFixtures, Files, FilesConfig} import ch.epfl.bluebrain.nexus.delta.plugins.storage.storages.model.{StorageRejection, StorageStatEntry, StorageType} import ch.epfl.bluebrain.nexus.delta.plugins.storage.storages.operations.remote.client.RemoteDiskStorageClient -import ch.epfl.bluebrain.nexus.delta.plugins.storage.storages.{StorageFixtures, Storages, StoragesConfig, StoragesStatistics, contexts => storageContexts, permissions => storagesPermissions} +import ch.epfl.bluebrain.nexus.delta.plugins.storage.storages.{contexts => storageContexts, permissions => storagesPermissions, StorageFixtures, Storages, StoragesConfig, StoragesStatistics} import ch.epfl.bluebrain.nexus.delta.rdf.IriOrBNode.Iri import ch.epfl.bluebrain.nexus.delta.rdf.RdfMediaTypes.`application/ld+json` import ch.epfl.bluebrain.nexus.delta.rdf.Vocabulary @@ -681,15 +681,17 @@ object FilesRoutesSpec extends CirceLiteral { createdBy: Subject, updatedBy: Subject )(implicit baseUri: BaseUri): Json = { - val self = ResourceUris("files", project, id).accessUri + val self = ResourceUris("files", project, id).accessUri val keywordsJson: Json = userMetadata match { case Some(meta) => Json.obj( - "keywords" -> JsonObject.fromIterable( - meta.keywords.map { - case (k, v) => k.value -> v.asJson - } - ).toJson + "keywords" -> JsonObject + .fromIterable( + meta.keywords.map { case (k, v) => + k.value -> v.asJson + } + ) + .toJson ) case None => Json.obj() } @@ -730,6 +732,6 @@ object FilesRoutesSpec extends CirceLiteral { } """ - mainJson deepMerge(keywordsJson) + mainJson deepMerge (keywordsJson) } }