WIP - Total
This commit is contained in:
parent
f3ee988bec
commit
ce15e2965c
6
.github/workflows/main.yml
vendored
6
.github/workflows/main.yml
vendored
@ -106,9 +106,9 @@ jobs:
|
|||||||
context: ./Encode
|
context: ./Encode
|
||||||
push: true
|
push: true
|
||||||
tags: |
|
tags: |
|
||||||
bskjon/media-processing--encode:latest
|
bskjon/media-processing-encode:latest
|
||||||
bskjon/media-processing--encode:${{ github.sha }}
|
bskjon/media-processing-encode:${{ github.sha }}
|
||||||
bskjon/media-processing--encode:${{ steps.docker-tag.outputs.tag }}
|
bskjon/media-processing-encode:${{ steps.docker-tag.outputs.tag }}
|
||||||
|
|
||||||
build-reader:
|
build-reader:
|
||||||
needs: build-commoncode
|
needs: build-commoncode
|
||||||
|
|||||||
@ -7,12 +7,26 @@ version = "1.0-SNAPSHOT"
|
|||||||
|
|
||||||
repositories {
|
repositories {
|
||||||
mavenCentral()
|
mavenCentral()
|
||||||
|
maven("https://jitpack.io")
|
||||||
|
maven {
|
||||||
|
url = uri("https://reposilite.iktdev.no/releases")
|
||||||
|
}
|
||||||
|
maven {
|
||||||
|
url = uri("https://reposilite.iktdev.no/snapshots")
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
dependencies {
|
dependencies {
|
||||||
implementation("com.github.pgreze:kotlin-process:1.3.1")
|
implementation("com.github.pgreze:kotlin-process:1.3.1")
|
||||||
implementation("io.github.microutils:kotlin-logging-jvm:2.0.11")
|
implementation("io.github.microutils:kotlin-logging-jvm:2.0.11")
|
||||||
|
|
||||||
|
implementation("no.iktdev.streamit.library:streamit-library-kafka:0.0.2-alpha63")
|
||||||
|
implementation("no.iktdev:exfl:0.0.12-SNAPSHOT")
|
||||||
|
|
||||||
|
implementation("com.google.code.gson:gson:2.8.9")
|
||||||
|
implementation("org.json:json:20230227")
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
testImplementation("junit:junit:4.13.2")
|
testImplementation("junit:junit:4.13.2")
|
||||||
testImplementation("org.junit.jupiter:junit-jupiter")
|
testImplementation("org.junit.jupiter:junit-jupiter")
|
||||||
|
|||||||
@ -4,6 +4,6 @@ import java.io.File
|
|||||||
|
|
||||||
object CommonConfig {
|
object CommonConfig {
|
||||||
var kafkaTopic: String = System.getenv("KAFKA_TOPIC") ?: "contentEvents"
|
var kafkaTopic: String = System.getenv("KAFKA_TOPIC") ?: "contentEvents"
|
||||||
var incomingContent: File? = if (!System.getenv("DIRECTORY_CONTENT_INCOMING").isNullOrBlank()) File(System.getenv("DIRECTORY_CONTENT_INCOMING")) else File("/src/input")
|
var incomingContent: File = if (!System.getenv("DIRECTORY_CONTENT_INCOMING").isNullOrBlank()) File(System.getenv("DIRECTORY_CONTENT_INCOMING")) else File("/src/input")
|
||||||
|
val outgoingContent: File = if (!System.getenv("DIRECTORY_CONTENT_OUTGOING").isNullOrBlank()) File(System.getenv("DIRECTORY_CONTENT_OUTGOING")) else File("/src/output")
|
||||||
}
|
}
|
||||||
@ -1,4 +1,4 @@
|
|||||||
package no.iktdev.streamit.content.reader
|
package no.iktdev.streamit.content.common
|
||||||
|
|
||||||
import no.iktdev.streamit.content.common.CommonConfig
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
import no.iktdev.streamit.library.kafka.KafkaEvents
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
@ -6,25 +6,34 @@ import no.iktdev.streamit.library.kafka.consumers.DefaultConsumer
|
|||||||
import no.iktdev.streamit.library.kafka.dto.Message
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
import no.iktdev.streamit.library.kafka.dto.Status
|
import no.iktdev.streamit.library.kafka.dto.Status
|
||||||
import no.iktdev.streamit.library.kafka.dto.StatusType
|
import no.iktdev.streamit.library.kafka.dto.StatusType
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
import no.iktdev.streamit.library.kafka.producer.DefaultProducer
|
import no.iktdev.streamit.library.kafka.producer.DefaultProducer
|
||||||
|
|
||||||
abstract class DefaultKafkaReader(val subId: String) {
|
abstract class DefaultKafkaReader(val subId: String) {
|
||||||
val messageProducer = DefaultProducer(CommonConfig.kafkaTopic)
|
val messageProducer = DefaultProducer(CommonConfig.kafkaTopic)
|
||||||
val defaultConsumer = DefaultConsumer(subId = subId)
|
val defaultConsumer = DefaultConsumer(subId = subId)
|
||||||
|
|
||||||
fun produceErrorMessage(baseMessage: Message, reason: String) {
|
abstract fun loadDeserializers(): Map<String, IMessageDataDeserialization<*>>
|
||||||
|
|
||||||
|
fun produceErrorMessage(event: KafkaEvents, baseMessage: Message, reason: String) {
|
||||||
val message = Message(
|
val message = Message(
|
||||||
referenceId = baseMessage.referenceId,
|
referenceId = baseMessage.referenceId,
|
||||||
actionType = baseMessage.actionType,
|
|
||||||
Status(statusType = StatusType.ERROR, message = reason)
|
Status(statusType = StatusType.ERROR, message = reason)
|
||||||
)
|
)
|
||||||
messageProducer.sendMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED.event, message)
|
messageProducer.sendMessage(event.event, message)
|
||||||
|
}
|
||||||
|
|
||||||
|
fun produceErrorMessage(event: KafkaEvents, referenceId: String, reason: String) {
|
||||||
|
val message = Message(
|
||||||
|
referenceId = referenceId,
|
||||||
|
Status(statusType = StatusType.ERROR, message = reason)
|
||||||
|
)
|
||||||
|
messageProducer.sendMessage(event.event, message)
|
||||||
}
|
}
|
||||||
|
|
||||||
fun produceMessage(event: KafkaEvents, baseMessage: Message, data: Any?) {
|
fun produceMessage(event: KafkaEvents, baseMessage: Message, data: Any?) {
|
||||||
val message = Message(
|
val message = Message(
|
||||||
referenceId = baseMessage.referenceId,
|
referenceId = baseMessage.referenceId,
|
||||||
actionType = baseMessage.actionType,
|
|
||||||
Status(statusType = if (data != null) StatusType.SUCCESS else StatusType.IGNORED),
|
Status(statusType = if (data != null) StatusType.SUCCESS else StatusType.IGNORED),
|
||||||
data = data
|
data = data
|
||||||
)
|
)
|
||||||
@ -0,0 +1,96 @@
|
|||||||
|
package no.iktdev.streamit.content.common
|
||||||
|
|
||||||
|
import no.iktdev.exfl.using
|
||||||
|
import java.io.File
|
||||||
|
import java.io.FileOutputStream
|
||||||
|
import java.net.HttpURLConnection
|
||||||
|
import java.net.URL
|
||||||
|
import kotlin.math.sign
|
||||||
|
|
||||||
|
open class Downloader(val url: String, val outDir: File, val baseName: String) {
|
||||||
|
protected val http: HttpURLConnection = openConnection()
|
||||||
|
private val BUFFER_SIZE = 4096
|
||||||
|
|
||||||
|
private fun openConnection(): HttpURLConnection {
|
||||||
|
try {
|
||||||
|
return URL(url).openConnection() as HttpURLConnection
|
||||||
|
} catch (e: Exception) {
|
||||||
|
e.printStackTrace()
|
||||||
|
throw BadAddressException("Provided url is either not provided (null) or is not a valid http url")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
protected fun getLength(): Int {
|
||||||
|
return http.contentLength
|
||||||
|
}
|
||||||
|
|
||||||
|
protected fun getProgress(read: Int, total: Int = getLength()): Int {
|
||||||
|
return ((read * 100) / total)
|
||||||
|
}
|
||||||
|
|
||||||
|
suspend fun download(): File? {
|
||||||
|
val extension = getExtension()
|
||||||
|
?: throw UnsupportedFormatException("Provided url does not contain a supported file extension")
|
||||||
|
val outFile = outDir.using("$baseName.$extension")
|
||||||
|
val inputStream = http.inputStream
|
||||||
|
val fos = FileOutputStream(outFile, false)
|
||||||
|
|
||||||
|
var totalBytesRead = 0
|
||||||
|
val buffer = ByteArray(BUFFER_SIZE)
|
||||||
|
inputStream.apply {
|
||||||
|
fos.use { fout ->
|
||||||
|
run {
|
||||||
|
var bytesRead = read(buffer)
|
||||||
|
while (bytesRead >= 0) {
|
||||||
|
fout.write(buffer, 0, bytesRead)
|
||||||
|
totalBytesRead += bytesRead
|
||||||
|
bytesRead = read(buffer)
|
||||||
|
// System.out.println(getProgress(totalBytesRead))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
inputStream.close()
|
||||||
|
fos.close()
|
||||||
|
return outFile
|
||||||
|
}
|
||||||
|
|
||||||
|
open fun getExtension(): String? {
|
||||||
|
val possiblyExtension = url.lastIndexOf(".") + 1
|
||||||
|
return if (possiblyExtension > 1) {
|
||||||
|
return url.toString().substring(possiblyExtension)
|
||||||
|
} else {
|
||||||
|
val mimeType = http.contentType ?: null
|
||||||
|
contentTypeToExtension()[mimeType]
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
open fun contentTypeToExtension(): Map<String, String> {
|
||||||
|
return mapOf(
|
||||||
|
"image/png" to "png",
|
||||||
|
"image/jpeg" to "jpg",
|
||||||
|
"image/webp" to "webp",
|
||||||
|
"image/bmp" to "bmp",
|
||||||
|
"image/tiff" to "tiff"
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class BadAddressException : java.lang.Exception {
|
||||||
|
constructor() : super() {}
|
||||||
|
constructor(message: String?) : super(message) {}
|
||||||
|
constructor(message: String?, cause: Throwable?) : super(message, cause) {}
|
||||||
|
}
|
||||||
|
|
||||||
|
class UnsupportedFormatException : Exception {
|
||||||
|
constructor() : super() {}
|
||||||
|
constructor(message: String?) : super(message) {}
|
||||||
|
constructor(message: String?, cause: Throwable?) : super(message, cause) {}
|
||||||
|
}
|
||||||
|
|
||||||
|
class InvalidFileException : Exception {
|
||||||
|
constructor() : super() {}
|
||||||
|
constructor(message: String?) : super(message) {}
|
||||||
|
constructor(message: String?, cause: Throwable?) : super(message, cause) {}
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,20 @@
|
|||||||
|
package no.iktdev.streamit.content.common
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
|
import no.iktdev.streamit.library.kafka.consumers.DefaultConsumer
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Status
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.StatusType
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.sequential.ISequentialMessageEvent
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.sequential.SequentialMessageListener
|
||||||
|
import no.iktdev.streamit.library.kafka.producer.DefaultProducer
|
||||||
|
|
||||||
|
abstract class SequentialKafkaReader(subId: String): DefaultKafkaReader(subId), ISequentialMessageEvent {
|
||||||
|
|
||||||
|
abstract val accept: KafkaEvents
|
||||||
|
abstract val subAccepts: List<KafkaEvents>
|
||||||
|
|
||||||
|
|
||||||
|
}
|
||||||
@ -7,9 +7,9 @@ import mu.KotlinLogging
|
|||||||
|
|
||||||
private val logger = KotlinLogging.logger {}
|
private val logger = KotlinLogging.logger {}
|
||||||
|
|
||||||
class Daemon(open val executable: String, val parameters: List<String>, val daemonInterface: IDaemon) {
|
open class Daemon(open val executable: String, val daemonInterface: IDaemon) {
|
||||||
var executor: ProcessResult? = null
|
var executor: ProcessResult? = null
|
||||||
suspend fun run(): Int {
|
open suspend fun run(parameters: List<String>): Int {
|
||||||
daemonInterface.onStarted()
|
daemonInterface.onStarted()
|
||||||
executor = process(executable, *parameters.toTypedArray(),
|
executor = process(executable, *parameters.toTypedArray(),
|
||||||
stdout = Redirect.CAPTURE,
|
stdout = Redirect.CAPTURE,
|
||||||
@ -20,7 +20,7 @@ class Daemon(open val executable: String, val parameters: List<String>, val daem
|
|||||||
val resultCode = executor?.resultCode ?: -1
|
val resultCode = executor?.resultCode ?: -1
|
||||||
if (resultCode == 0) {
|
if (resultCode == 0) {
|
||||||
daemonInterface.onEnded()
|
daemonInterface.onEnded()
|
||||||
} else daemonInterface.onError()
|
} else daemonInterface.onError(resultCode)
|
||||||
return resultCode
|
return resultCode
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -8,6 +8,6 @@ interface IDaemon {
|
|||||||
|
|
||||||
fun onEnded() {}
|
fun onEnded() {}
|
||||||
|
|
||||||
fun onError()
|
fun onError(code: Int)
|
||||||
|
|
||||||
}
|
}
|
||||||
@ -0,0 +1,11 @@
|
|||||||
|
package no.iktdev.streamit.content.common.deserializers
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.dto.ContentOutName
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
|
||||||
|
class ContentOutNameDeserializer: IMessageDataDeserialization<ContentOutName> {
|
||||||
|
override fun deserialize(incomingMessage: Message): ContentOutName? {
|
||||||
|
return incomingMessage.dataAs(ContentOutName::class.java)
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,36 @@
|
|||||||
|
package no.iktdev.streamit.content.common.deserializers
|
||||||
|
|
||||||
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
|
||||||
|
class DeserializerRegistry {
|
||||||
|
companion object {
|
||||||
|
private val _registry = mutableMapOf<KafkaEvents, IMessageDataDeserialization<*>>(
|
||||||
|
KafkaEvents.EVENT_READER_RECEIVED_FILE to FileResultDeserializer(),
|
||||||
|
KafkaEvents.EVENT_METADATA_OBTAINED to MetadataResultDeserializer(),
|
||||||
|
KafkaEvents.EVENT_READER_DETERMINED_SERIE to EpisodeInfoDeserializer(),
|
||||||
|
KafkaEvents.EVENT_READER_DETERMINED_MOVIE to MovieInfoDeserializer(),
|
||||||
|
|
||||||
|
KafkaEvents.EVENT_READER_ENCODE_GENERATED_VIDEO to EncodeWorkDeserializer(),
|
||||||
|
KafkaEvents.EVENT_ENCODER_ENDED_VIDEO_FILE to EncodeWorkDeserializer(),
|
||||||
|
KafkaEvents.EVENT_READER_ENCODE_GENERATED_SUBTITLE to ExtractWorkDeserializer(),
|
||||||
|
KafkaEvents.EVENT_ENCODER_ENDED_SUBTITLE_FILE to ExtractWorkDeserializer(),
|
||||||
|
|
||||||
|
)
|
||||||
|
fun getRegistry(): Map<KafkaEvents, IMessageDataDeserialization<*>> = _registry.toMap()
|
||||||
|
fun getEventToDeserializer(vararg keys: KafkaEvents): Map<String, IMessageDataDeserialization<*>> {
|
||||||
|
val missingFields = keys.filter { !getRegistry().keys.contains(it) }
|
||||||
|
|
||||||
|
if (missingFields.isNotEmpty()) {
|
||||||
|
throw MissingDeserializerException("Missing deserializers for: ${missingFields.joinToString(", ")}")
|
||||||
|
}
|
||||||
|
return getRegistry().filter { keys.contains(it.key) }.map { it.key.event to it.value }.toMap()
|
||||||
|
}
|
||||||
|
fun addDeserializer(key: KafkaEvents, deserializer: IMessageDataDeserialization<*>) {
|
||||||
|
_registry[key] = deserializer
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class MissingDeserializerException(override val message: String): RuntimeException()
|
||||||
@ -0,0 +1,11 @@
|
|||||||
|
package no.iktdev.streamit.content.common.deserializers
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.EncodeWork
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
|
||||||
|
class EncodeWorkDeserializer: IMessageDataDeserialization<EncodeWork> {
|
||||||
|
override fun deserialize(incomingMessage: Message): EncodeWork? {
|
||||||
|
return incomingMessage.dataAs(EncodeWork::class.java)
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,11 @@
|
|||||||
|
package no.iktdev.streamit.content.common.deserializers
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.EpisodeInfo
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
|
||||||
|
class EpisodeInfoDeserializer: IMessageDataDeserialization<EpisodeInfo> {
|
||||||
|
override fun deserialize(incomingMessage: Message): EpisodeInfo? {
|
||||||
|
return incomingMessage.dataAs(EpisodeInfo::class.java)
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,12 @@
|
|||||||
|
package no.iktdev.streamit.content.common.deserializers
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.EncodeWork
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.ExtractWork
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
|
||||||
|
class ExtractWorkDeserializer: IMessageDataDeserialization<ExtractWork> {
|
||||||
|
override fun deserialize(incomingMessage: Message): ExtractWork? {
|
||||||
|
return incomingMessage.dataAs(ExtractWork::class.java)
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,13 @@
|
|||||||
|
package no.iktdev.streamit.content.common.deserializers
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.FileResult
|
||||||
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.StatusType
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
|
||||||
|
class FileResultDeserializer: IMessageDataDeserialization<FileResult> {
|
||||||
|
override fun deserialize(incomingMessage: Message): FileResult? {
|
||||||
|
return incomingMessage.dataAs(FileResult::class.java)
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,47 @@
|
|||||||
|
package no.iktdev.streamit.content.common.deserializers
|
||||||
|
|
||||||
|
import com.google.gson.Gson
|
||||||
|
import com.google.gson.JsonObject
|
||||||
|
import no.iktdev.streamit.content.common.streams.AudioStream
|
||||||
|
import no.iktdev.streamit.content.common.streams.MediaStreams
|
||||||
|
import no.iktdev.streamit.content.common.streams.SubtitleStream
|
||||||
|
import no.iktdev.streamit.content.common.streams.VideoStream
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.StatusType
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
|
||||||
|
class MediaStreamsDeserializer: IMessageDataDeserialization<MediaStreams> {
|
||||||
|
override fun deserialize(incomingMessage: Message): MediaStreams? {
|
||||||
|
return try {
|
||||||
|
val gson = Gson()
|
||||||
|
val jsonObject = if (incomingMessage.data is String) {
|
||||||
|
gson.fromJson(incomingMessage.data as String, JsonObject::class.java)
|
||||||
|
} else {
|
||||||
|
gson.fromJson(incomingMessage.dataAsJson(), JsonObject::class.java)
|
||||||
|
}
|
||||||
|
|
||||||
|
val streamsJsonArray = jsonObject.getAsJsonArray("streams")
|
||||||
|
|
||||||
|
val rstreams = streamsJsonArray.mapNotNull { streamJson ->
|
||||||
|
val streamObject = streamJson.asJsonObject
|
||||||
|
|
||||||
|
val codecType = streamObject.get("codec_type").asString
|
||||||
|
if (streamObject.has("codec_name") && streamObject.get("codec_name").asString == "mjpeg") {
|
||||||
|
null
|
||||||
|
} else {
|
||||||
|
when (codecType) {
|
||||||
|
"video" -> gson.fromJson(streamObject, VideoStream::class.java)
|
||||||
|
"audio" -> gson.fromJson(streamObject, AudioStream::class.java)
|
||||||
|
"subtitle" -> gson.fromJson(streamObject, SubtitleStream::class.java)
|
||||||
|
else -> null //throw IllegalArgumentException("Unknown stream type: $codecType")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return MediaStreams(rstreams)
|
||||||
|
} catch (e: Exception) {
|
||||||
|
e.printStackTrace()
|
||||||
|
null
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,11 @@
|
|||||||
|
package no.iktdev.streamit.content.common.deserializers
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.dto.Metadata
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
|
||||||
|
class MetadataResultDeserializer: IMessageDataDeserialization<Metadata> {
|
||||||
|
override fun deserialize(incomingMessage: Message): Metadata? {
|
||||||
|
return incomingMessage.dataAs(Metadata::class.java)
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,11 @@
|
|||||||
|
package no.iktdev.streamit.content.common.deserializers
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.MovieInfo
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
|
||||||
|
class MovieInfoDeserializer: IMessageDataDeserialization<MovieInfo> {
|
||||||
|
override fun deserialize(incomingMessage: Message): MovieInfo? {
|
||||||
|
return incomingMessage.dataAs(MovieInfo::class.java)
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,9 @@
|
|||||||
|
package no.iktdev.streamit.content.common.dto.reader
|
||||||
|
|
||||||
|
data class EpisodeInfo(
|
||||||
|
val title: String,
|
||||||
|
val episode: Int,
|
||||||
|
val season: Int,
|
||||||
|
val episodeTitle: String?,
|
||||||
|
override val fullName: String
|
||||||
|
): VideoInfo(fullName)
|
||||||
@ -0,0 +1,7 @@
|
|||||||
|
package no.iktdev.streamit.content.common.dto.reader
|
||||||
|
|
||||||
|
data class FileResult(
|
||||||
|
val file: String,
|
||||||
|
val title: String = "",
|
||||||
|
val sanitizedName: String = ""
|
||||||
|
)
|
||||||
@ -0,0 +1,6 @@
|
|||||||
|
package no.iktdev.streamit.content.common.dto.reader
|
||||||
|
|
||||||
|
data class MovieInfo(
|
||||||
|
val title: String,
|
||||||
|
override val fullName: String
|
||||||
|
) : VideoInfo(fullName)
|
||||||
@ -0,0 +1,4 @@
|
|||||||
|
package no.iktdev.streamit.content.common.dto.reader
|
||||||
|
|
||||||
|
class SubtitleInfo {
|
||||||
|
}
|
||||||
@ -0,0 +1,5 @@
|
|||||||
|
package no.iktdev.streamit.content.common.dto.reader
|
||||||
|
|
||||||
|
abstract class VideoInfo(
|
||||||
|
@Transient open val fullName: String
|
||||||
|
)
|
||||||
@ -0,0 +1,8 @@
|
|||||||
|
package no.iktdev.streamit.content.common.dto.reader.work
|
||||||
|
|
||||||
|
data class ConvertWork(
|
||||||
|
override val collection: String,
|
||||||
|
val language: String,
|
||||||
|
override val inFile: String,
|
||||||
|
override val outFile: String,
|
||||||
|
) : WorkBase(collection = collection, inFile = inFile, outFile = outFile)
|
||||||
@ -0,0 +1,8 @@
|
|||||||
|
package no.iktdev.streamit.content.common.dto.reader.work
|
||||||
|
|
||||||
|
data class EncodeWork(
|
||||||
|
override val collection: String,
|
||||||
|
override val inFile: String,
|
||||||
|
override val outFile: String,
|
||||||
|
val arguments: List<String>
|
||||||
|
) : WorkBase(collection = collection, inFile = inFile, outFile = outFile)
|
||||||
@ -0,0 +1,10 @@
|
|||||||
|
package no.iktdev.streamit.content.common.dto.reader.work
|
||||||
|
|
||||||
|
data class ExtractWork(
|
||||||
|
override val collection: String,
|
||||||
|
val language: String,
|
||||||
|
override val inFile: String,
|
||||||
|
val arguments: List<String>,
|
||||||
|
override val outFile: String,
|
||||||
|
var produceConvertEvent: Boolean = true
|
||||||
|
) : WorkBase(collection = collection, inFile = inFile, outFile = outFile)
|
||||||
@ -0,0 +1,10 @@
|
|||||||
|
package no.iktdev.streamit.content.common.dto.reader.work
|
||||||
|
|
||||||
|
import java.util.UUID
|
||||||
|
|
||||||
|
abstract class WorkBase(
|
||||||
|
@Transient open val workId: String = UUID.randomUUID().toString(),
|
||||||
|
@Transient open val collection: String,
|
||||||
|
@Transient open val inFile: String,
|
||||||
|
@Transient open val outFile: String
|
||||||
|
)
|
||||||
@ -0,0 +1,22 @@
|
|||||||
|
package no.iktdev.streamit.content.common.streams
|
||||||
|
|
||||||
|
class SubtitleStreamSelector(val streams: List<SubtitleStream>) {
|
||||||
|
|
||||||
|
fun getDesiredStreams(): List<SubtitleStream> {
|
||||||
|
val codecFiltered = streams.filter { getFormatToCodec(it.codec_name) != null }
|
||||||
|
// TODO: Expand and remove stuff like sign and songs etc..
|
||||||
|
return codecFiltered
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
fun getFormatToCodec(codecName: String): String? {
|
||||||
|
return when(codecName) {
|
||||||
|
"ass" -> "ass"
|
||||||
|
"subrip" -> "srt"
|
||||||
|
"webvtt", "vtt" -> "vtt"
|
||||||
|
"smi" -> "smi"
|
||||||
|
"hdmv_pgs_subtitle" -> null
|
||||||
|
else -> null
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
BIN
Convert/gradle/wrapper/gradle-wrapper.jar
vendored
Normal file
BIN
Convert/gradle/wrapper/gradle-wrapper.jar
vendored
Normal file
Binary file not shown.
6
Convert/gradle/wrapper/gradle-wrapper.properties
vendored
Normal file
6
Convert/gradle/wrapper/gradle-wrapper.properties
vendored
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
#Sun Jul 23 01:48:17 CEST 2023
|
||||||
|
distributionBase=GRADLE_USER_HOME
|
||||||
|
distributionPath=wrapper/dists
|
||||||
|
distributionUrl=https\://services.gradle.org/distributions/gradle-7.6.1-bin.zip
|
||||||
|
zipStoreBase=GRADLE_USER_HOME
|
||||||
|
zipStorePath=wrapper/dists
|
||||||
234
Convert/gradlew
vendored
Normal file
234
Convert/gradlew
vendored
Normal file
@ -0,0 +1,234 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
#
|
||||||
|
# Copyright © 2015-2021 the original authors.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# https://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
##############################################################################
|
||||||
|
#
|
||||||
|
# Gradle start up script for POSIX generated by Gradle.
|
||||||
|
#
|
||||||
|
# Important for running:
|
||||||
|
#
|
||||||
|
# (1) You need a POSIX-compliant shell to run this script. If your /bin/sh is
|
||||||
|
# noncompliant, but you have some other compliant shell such as ksh or
|
||||||
|
# bash, then to run this script, type that shell name before the whole
|
||||||
|
# command line, like:
|
||||||
|
#
|
||||||
|
# ksh Gradle
|
||||||
|
#
|
||||||
|
# Busybox and similar reduced shells will NOT work, because this script
|
||||||
|
# requires all of these POSIX shell features:
|
||||||
|
# * functions;
|
||||||
|
# * expansions «$var», «${var}», «${var:-default}», «${var+SET}»,
|
||||||
|
# «${var#prefix}», «${var%suffix}», and «$( cmd )»;
|
||||||
|
# * compound commands having a testable exit status, especially «case»;
|
||||||
|
# * various built-in commands including «command», «set», and «ulimit».
|
||||||
|
#
|
||||||
|
# Important for patching:
|
||||||
|
#
|
||||||
|
# (2) This script targets any POSIX shell, so it avoids extensions provided
|
||||||
|
# by Bash, Ksh, etc; in particular arrays are avoided.
|
||||||
|
#
|
||||||
|
# The "traditional" practice of packing multiple parameters into a
|
||||||
|
# space-separated string is a well documented source of bugs and security
|
||||||
|
# problems, so this is (mostly) avoided, by progressively accumulating
|
||||||
|
# options in "$@", and eventually passing that to Java.
|
||||||
|
#
|
||||||
|
# Where the inherited environment variables (DEFAULT_JVM_OPTS, JAVA_OPTS,
|
||||||
|
# and GRADLE_OPTS) rely on word-splitting, this is performed explicitly;
|
||||||
|
# see the in-line comments for details.
|
||||||
|
#
|
||||||
|
# There are tweaks for specific operating systems such as AIX, CygWin,
|
||||||
|
# Darwin, MinGW, and NonStop.
|
||||||
|
#
|
||||||
|
# (3) This script is generated from the Groovy template
|
||||||
|
# https://github.com/gradle/gradle/blob/master/subprojects/plugins/src/main/resources/org/gradle/api/internal/plugins/unixStartScript.txt
|
||||||
|
# within the Gradle project.
|
||||||
|
#
|
||||||
|
# You can find Gradle at https://github.com/gradle/gradle/.
|
||||||
|
#
|
||||||
|
##############################################################################
|
||||||
|
|
||||||
|
# Attempt to set APP_HOME
|
||||||
|
|
||||||
|
# Resolve links: $0 may be a link
|
||||||
|
app_path=$0
|
||||||
|
|
||||||
|
# Need this for daisy-chained symlinks.
|
||||||
|
while
|
||||||
|
APP_HOME=${app_path%"${app_path##*/}"} # leaves a trailing /; empty if no leading path
|
||||||
|
[ -h "$app_path" ]
|
||||||
|
do
|
||||||
|
ls=$( ls -ld "$app_path" )
|
||||||
|
link=${ls#*' -> '}
|
||||||
|
case $link in #(
|
||||||
|
/*) app_path=$link ;; #(
|
||||||
|
*) app_path=$APP_HOME$link ;;
|
||||||
|
esac
|
||||||
|
done
|
||||||
|
|
||||||
|
APP_HOME=$( cd "${APP_HOME:-./}" && pwd -P ) || exit
|
||||||
|
|
||||||
|
APP_NAME="Gradle"
|
||||||
|
APP_BASE_NAME=${0##*/}
|
||||||
|
|
||||||
|
# Add default JVM options here. You can also use JAVA_OPTS and GRADLE_OPTS to pass JVM options to this script.
|
||||||
|
DEFAULT_JVM_OPTS='"-Xmx64m" "-Xms64m"'
|
||||||
|
|
||||||
|
# Use the maximum available, or set MAX_FD != -1 to use that value.
|
||||||
|
MAX_FD=maximum
|
||||||
|
|
||||||
|
warn () {
|
||||||
|
echo "$*"
|
||||||
|
} >&2
|
||||||
|
|
||||||
|
die () {
|
||||||
|
echo
|
||||||
|
echo "$*"
|
||||||
|
echo
|
||||||
|
exit 1
|
||||||
|
} >&2
|
||||||
|
|
||||||
|
# OS specific support (must be 'true' or 'false').
|
||||||
|
cygwin=false
|
||||||
|
msys=false
|
||||||
|
darwin=false
|
||||||
|
nonstop=false
|
||||||
|
case "$( uname )" in #(
|
||||||
|
CYGWIN* ) cygwin=true ;; #(
|
||||||
|
Darwin* ) darwin=true ;; #(
|
||||||
|
MSYS* | MINGW* ) msys=true ;; #(
|
||||||
|
NONSTOP* ) nonstop=true ;;
|
||||||
|
esac
|
||||||
|
|
||||||
|
CLASSPATH=$APP_HOME/gradle/wrapper/gradle-wrapper.jar
|
||||||
|
|
||||||
|
|
||||||
|
# Determine the Java command to use to start the JVM.
|
||||||
|
if [ -n "$JAVA_HOME" ] ; then
|
||||||
|
if [ -x "$JAVA_HOME/jre/sh/java" ] ; then
|
||||||
|
# IBM's JDK on AIX uses strange locations for the executables
|
||||||
|
JAVACMD=$JAVA_HOME/jre/sh/java
|
||||||
|
else
|
||||||
|
JAVACMD=$JAVA_HOME/bin/java
|
||||||
|
fi
|
||||||
|
if [ ! -x "$JAVACMD" ] ; then
|
||||||
|
die "ERROR: JAVA_HOME is set to an invalid directory: $JAVA_HOME
|
||||||
|
|
||||||
|
Please set the JAVA_HOME variable in your environment to match the
|
||||||
|
location of your Java installation."
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
JAVACMD=java
|
||||||
|
which java >/dev/null 2>&1 || die "ERROR: JAVA_HOME is not set and no 'java' command could be found in your PATH.
|
||||||
|
|
||||||
|
Please set the JAVA_HOME variable in your environment to match the
|
||||||
|
location of your Java installation."
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Increase the maximum file descriptors if we can.
|
||||||
|
if ! "$cygwin" && ! "$darwin" && ! "$nonstop" ; then
|
||||||
|
case $MAX_FD in #(
|
||||||
|
max*)
|
||||||
|
MAX_FD=$( ulimit -H -n ) ||
|
||||||
|
warn "Could not query maximum file descriptor limit"
|
||||||
|
esac
|
||||||
|
case $MAX_FD in #(
|
||||||
|
'' | soft) :;; #(
|
||||||
|
*)
|
||||||
|
ulimit -n "$MAX_FD" ||
|
||||||
|
warn "Could not set maximum file descriptor limit to $MAX_FD"
|
||||||
|
esac
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Collect all arguments for the java command, stacking in reverse order:
|
||||||
|
# * args from the command line
|
||||||
|
# * the main class name
|
||||||
|
# * -classpath
|
||||||
|
# * -D...appname settings
|
||||||
|
# * --module-path (only if needed)
|
||||||
|
# * DEFAULT_JVM_OPTS, JAVA_OPTS, and GRADLE_OPTS environment variables.
|
||||||
|
|
||||||
|
# For Cygwin or MSYS, switch paths to Windows format before running java
|
||||||
|
if "$cygwin" || "$msys" ; then
|
||||||
|
APP_HOME=$( cygpath --path --mixed "$APP_HOME" )
|
||||||
|
CLASSPATH=$( cygpath --path --mixed "$CLASSPATH" )
|
||||||
|
|
||||||
|
JAVACMD=$( cygpath --unix "$JAVACMD" )
|
||||||
|
|
||||||
|
# Now convert the arguments - kludge to limit ourselves to /bin/sh
|
||||||
|
for arg do
|
||||||
|
if
|
||||||
|
case $arg in #(
|
||||||
|
-*) false ;; # don't mess with options #(
|
||||||
|
/?*) t=${arg#/} t=/${t%%/*} # looks like a POSIX filepath
|
||||||
|
[ -e "$t" ] ;; #(
|
||||||
|
*) false ;;
|
||||||
|
esac
|
||||||
|
then
|
||||||
|
arg=$( cygpath --path --ignore --mixed "$arg" )
|
||||||
|
fi
|
||||||
|
# Roll the args list around exactly as many times as the number of
|
||||||
|
# args, so each arg winds up back in the position where it started, but
|
||||||
|
# possibly modified.
|
||||||
|
#
|
||||||
|
# NB: a `for` loop captures its iteration list before it begins, so
|
||||||
|
# changing the positional parameters here affects neither the number of
|
||||||
|
# iterations, nor the values presented in `arg`.
|
||||||
|
shift # remove old arg
|
||||||
|
set -- "$@" "$arg" # push replacement arg
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Collect all arguments for the java command;
|
||||||
|
# * $DEFAULT_JVM_OPTS, $JAVA_OPTS, and $GRADLE_OPTS can contain fragments of
|
||||||
|
# shell script including quotes and variable substitutions, so put them in
|
||||||
|
# double quotes to make sure that they get re-expanded; and
|
||||||
|
# * put everything else in single quotes, so that it's not re-expanded.
|
||||||
|
|
||||||
|
set -- \
|
||||||
|
"-Dorg.gradle.appname=$APP_BASE_NAME" \
|
||||||
|
-classpath "$CLASSPATH" \
|
||||||
|
org.gradle.wrapper.GradleWrapperMain \
|
||||||
|
"$@"
|
||||||
|
|
||||||
|
# Use "xargs" to parse quoted args.
|
||||||
|
#
|
||||||
|
# With -n1 it outputs one arg per line, with the quotes and backslashes removed.
|
||||||
|
#
|
||||||
|
# In Bash we could simply go:
|
||||||
|
#
|
||||||
|
# readarray ARGS < <( xargs -n1 <<<"$var" ) &&
|
||||||
|
# set -- "${ARGS[@]}" "$@"
|
||||||
|
#
|
||||||
|
# but POSIX shell has neither arrays nor command substitution, so instead we
|
||||||
|
# post-process each arg (as a line of input to sed) to backslash-escape any
|
||||||
|
# character that might be a shell metacharacter, then use eval to reverse
|
||||||
|
# that process (while maintaining the separation between arguments), and wrap
|
||||||
|
# the whole thing up as a single "set" statement.
|
||||||
|
#
|
||||||
|
# This will of course break if any of these variables contains a newline or
|
||||||
|
# an unmatched quote.
|
||||||
|
#
|
||||||
|
|
||||||
|
eval "set -- $(
|
||||||
|
printf '%s\n' "$DEFAULT_JVM_OPTS $JAVA_OPTS $GRADLE_OPTS" |
|
||||||
|
xargs -n1 |
|
||||||
|
sed ' s~[^-[:alnum:]+,./:=@_]~\\&~g; ' |
|
||||||
|
tr '\n' ' '
|
||||||
|
)" '"$@"'
|
||||||
|
|
||||||
|
exec "$JAVACMD" "$@"
|
||||||
89
Convert/gradlew.bat
vendored
Normal file
89
Convert/gradlew.bat
vendored
Normal file
@ -0,0 +1,89 @@
|
|||||||
|
@rem
|
||||||
|
@rem Copyright 2015 the original author or authors.
|
||||||
|
@rem
|
||||||
|
@rem Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
@rem you may not use this file except in compliance with the License.
|
||||||
|
@rem You may obtain a copy of the License at
|
||||||
|
@rem
|
||||||
|
@rem https://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
@rem
|
||||||
|
@rem Unless required by applicable law or agreed to in writing, software
|
||||||
|
@rem distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
@rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
@rem See the License for the specific language governing permissions and
|
||||||
|
@rem limitations under the License.
|
||||||
|
@rem
|
||||||
|
|
||||||
|
@if "%DEBUG%" == "" @echo off
|
||||||
|
@rem ##########################################################################
|
||||||
|
@rem
|
||||||
|
@rem Gradle startup script for Windows
|
||||||
|
@rem
|
||||||
|
@rem ##########################################################################
|
||||||
|
|
||||||
|
@rem Set local scope for the variables with windows NT shell
|
||||||
|
if "%OS%"=="Windows_NT" setlocal
|
||||||
|
|
||||||
|
set DIRNAME=%~dp0
|
||||||
|
if "%DIRNAME%" == "" set DIRNAME=.
|
||||||
|
set APP_BASE_NAME=%~n0
|
||||||
|
set APP_HOME=%DIRNAME%
|
||||||
|
|
||||||
|
@rem Resolve any "." and ".." in APP_HOME to make it shorter.
|
||||||
|
for %%i in ("%APP_HOME%") do set APP_HOME=%%~fi
|
||||||
|
|
||||||
|
@rem Add default JVM options here. You can also use JAVA_OPTS and GRADLE_OPTS to pass JVM options to this script.
|
||||||
|
set DEFAULT_JVM_OPTS="-Xmx64m" "-Xms64m"
|
||||||
|
|
||||||
|
@rem Find java.exe
|
||||||
|
if defined JAVA_HOME goto findJavaFromJavaHome
|
||||||
|
|
||||||
|
set JAVA_EXE=java.exe
|
||||||
|
%JAVA_EXE% -version >NUL 2>&1
|
||||||
|
if "%ERRORLEVEL%" == "0" goto execute
|
||||||
|
|
||||||
|
echo.
|
||||||
|
echo ERROR: JAVA_HOME is not set and no 'java' command could be found in your PATH.
|
||||||
|
echo.
|
||||||
|
echo Please set the JAVA_HOME variable in your environment to match the
|
||||||
|
echo location of your Java installation.
|
||||||
|
|
||||||
|
goto fail
|
||||||
|
|
||||||
|
:findJavaFromJavaHome
|
||||||
|
set JAVA_HOME=%JAVA_HOME:"=%
|
||||||
|
set JAVA_EXE=%JAVA_HOME%/bin/java.exe
|
||||||
|
|
||||||
|
if exist "%JAVA_EXE%" goto execute
|
||||||
|
|
||||||
|
echo.
|
||||||
|
echo ERROR: JAVA_HOME is set to an invalid directory: %JAVA_HOME%
|
||||||
|
echo.
|
||||||
|
echo Please set the JAVA_HOME variable in your environment to match the
|
||||||
|
echo location of your Java installation.
|
||||||
|
|
||||||
|
goto fail
|
||||||
|
|
||||||
|
:execute
|
||||||
|
@rem Setup the command line
|
||||||
|
|
||||||
|
set CLASSPATH=%APP_HOME%\gradle\wrapper\gradle-wrapper.jar
|
||||||
|
|
||||||
|
|
||||||
|
@rem Execute Gradle
|
||||||
|
"%JAVA_EXE%" %DEFAULT_JVM_OPTS% %JAVA_OPTS% %GRADLE_OPTS% "-Dorg.gradle.appname=%APP_BASE_NAME%" -classpath "%CLASSPATH%" org.gradle.wrapper.GradleWrapperMain %*
|
||||||
|
|
||||||
|
:end
|
||||||
|
@rem End local scope for the variables with windows NT shell
|
||||||
|
if "%ERRORLEVEL%"=="0" goto mainEnd
|
||||||
|
|
||||||
|
:fail
|
||||||
|
rem Set variable GRADLE_EXIT_CONSOLE if you need the _script_ return code instead of
|
||||||
|
rem the _cmd.exe /c_ return code!
|
||||||
|
if not "" == "%GRADLE_EXIT_CONSOLE%" exit 1
|
||||||
|
exit /b 1
|
||||||
|
|
||||||
|
:mainEnd
|
||||||
|
if "%OS%"=="Windows_NT" endlocal
|
||||||
|
|
||||||
|
:omega
|
||||||
@ -0,0 +1,19 @@
|
|||||||
|
package no.iktdev.streamit.content.convert
|
||||||
|
|
||||||
|
import mu.KotlinLogging
|
||||||
|
import org.springframework.boot.autoconfigure.SpringBootApplication
|
||||||
|
import org.springframework.boot.runApplication
|
||||||
|
import org.springframework.context.ApplicationContext
|
||||||
|
|
||||||
|
@SpringBootApplication
|
||||||
|
class ConvertApplication
|
||||||
|
|
||||||
|
private var context: ApplicationContext? = null
|
||||||
|
@Suppress("unused")
|
||||||
|
fun getContext(): ApplicationContext? {
|
||||||
|
return context
|
||||||
|
}
|
||||||
|
fun main(args: Array<String>) {
|
||||||
|
context = runApplication<ConvertApplication>(*args)
|
||||||
|
}
|
||||||
|
private val logger = KotlinLogging.logger {}
|
||||||
@ -0,0 +1,4 @@
|
|||||||
|
package no.iktdev.streamit.content.convert
|
||||||
|
|
||||||
|
class ConvertEnv {
|
||||||
|
}
|
||||||
@ -0,0 +1,19 @@
|
|||||||
|
package no.iktdev.streamit.content.convert.kafka
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.DefaultKafkaReader
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.SimpleMessageListener
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
import org.springframework.stereotype.Service
|
||||||
|
|
||||||
|
@Service
|
||||||
|
class SubtitleConsumer: DefaultKafkaReader("convertHandler") {
|
||||||
|
|
||||||
|
/*init {
|
||||||
|
object: SimpleMessageListener(topic =b )
|
||||||
|
}*/
|
||||||
|
|
||||||
|
override fun loadDeserializers(): Map<String, IMessageDataDeserialization<*>> {
|
||||||
|
TODO("Not yet implemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
4
Encode/Dockerfile
Normal file
4
Encode/Dockerfile
Normal file
@ -0,0 +1,4 @@
|
|||||||
|
FROM bskjon/debian-azuljava17-ffmpeg:latest
|
||||||
|
EXPOSE 8080
|
||||||
|
|
||||||
|
COPY ./build/libs/encode.jar /usr/share/app/app.jar
|
||||||
@ -1,5 +1,8 @@
|
|||||||
plugins {
|
plugins {
|
||||||
kotlin("jvm") version "1.8.21"
|
kotlin("jvm") version "1.8.21"
|
||||||
|
id("org.springframework.boot") version "2.5.5"
|
||||||
|
id("io.spring.dependency-management") version "1.0.11.RELEASE"
|
||||||
|
kotlin("plugin.spring") version "1.5.31"
|
||||||
}
|
}
|
||||||
|
|
||||||
group = "no.iktdev.streamit.content"
|
group = "no.iktdev.streamit.content"
|
||||||
@ -7,9 +10,32 @@ version = "1.0-SNAPSHOT"
|
|||||||
|
|
||||||
repositories {
|
repositories {
|
||||||
mavenCentral()
|
mavenCentral()
|
||||||
|
maven("https://jitpack.io")
|
||||||
|
maven {
|
||||||
|
url = uri("https://reposilite.iktdev.no/releases")
|
||||||
|
}
|
||||||
|
maven {
|
||||||
|
url = uri("https://reposilite.iktdev.no/snapshots")
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
dependencies {
|
dependencies {
|
||||||
|
implementation(project(":CommonCode"))
|
||||||
|
|
||||||
|
implementation("no.iktdev.streamit.library:streamit-library-kafka:0.0.2-alpha63")
|
||||||
|
implementation("no.iktdev:exfl:0.0.12-SNAPSHOT")
|
||||||
|
|
||||||
|
implementation("org.jetbrains.kotlinx:kotlinx-coroutines-core:1.7.1")
|
||||||
|
|
||||||
|
implementation("com.github.pgreze:kotlin-process:1.3.1")
|
||||||
|
implementation("io.github.microutils:kotlin-logging-jvm:2.0.11")
|
||||||
|
|
||||||
|
implementation("com.google.code.gson:gson:2.8.9")
|
||||||
|
|
||||||
|
implementation("org.springframework.boot:spring-boot-starter-web")
|
||||||
|
implementation("org.springframework.boot:spring-boot-starter:2.7.0")
|
||||||
|
implementation("org.springframework.kafka:spring-kafka:2.8.5")
|
||||||
|
|
||||||
|
|
||||||
testImplementation(platform("org.junit:junit-bom:5.9.1"))
|
testImplementation(platform("org.junit:junit-bom:5.9.1"))
|
||||||
testImplementation("org.junit.jupiter:junit-jupiter")
|
testImplementation("org.junit.jupiter:junit-jupiter")
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,2 +1,4 @@
|
|||||||
rootProject.name = "Encode"
|
rootProject.name = "Encode"
|
||||||
|
|
||||||
|
include(":CommonCode")
|
||||||
|
project(":CommonCode").projectDir = File("../CommonCode")
|
||||||
7
Encode/src/main/kotlin/EncodeEnv.kt
Normal file
7
Encode/src/main/kotlin/EncodeEnv.kt
Normal file
@ -0,0 +1,7 @@
|
|||||||
|
class EncodeEnv {
|
||||||
|
companion object {
|
||||||
|
val ffmpeg: String = System.getenv("SUPPORTING_EXECUTABLE_FFMPEG") ?: "ffmpeg"
|
||||||
|
val allowOverwrite = System.getenv("ALLOW_OVERWRITE").toBoolean() ?: false
|
||||||
|
val maxRunners: Int = System.getenv("SIMULTANEOUS_ENCODE_RUNNERS").toIntOrNull() ?: 1
|
||||||
|
}
|
||||||
|
}
|
||||||
28
Encode/src/main/kotlin/EncoderApplication.kt
Normal file
28
Encode/src/main/kotlin/EncoderApplication.kt
Normal file
@ -0,0 +1,28 @@
|
|||||||
|
import mu.KotlinLogging
|
||||||
|
import no.iktdev.exfl.observable.ObservableMap
|
||||||
|
import org.springframework.boot.autoconfigure.SpringBootApplication
|
||||||
|
import org.springframework.boot.runApplication
|
||||||
|
import org.springframework.context.ApplicationContext
|
||||||
|
|
||||||
|
@SpringBootApplication
|
||||||
|
class EncoderApplication
|
||||||
|
|
||||||
|
private var context: ApplicationContext? = null
|
||||||
|
|
||||||
|
@Suppress("unused")
|
||||||
|
fun getContext(): ApplicationContext? {
|
||||||
|
return context
|
||||||
|
}
|
||||||
|
fun main(args: Array<String>) {
|
||||||
|
context = runApplication<EncoderApplication>(*args)
|
||||||
|
}
|
||||||
|
private val logger = KotlinLogging.logger {}
|
||||||
|
|
||||||
|
/*val progress = ObservableMap<String, EncodeInformation>().also {
|
||||||
|
it.addListener(object: ObservableMap.Listener<String, EncodeInformation> {
|
||||||
|
override fun onPut(key: String, value: EncodeInformation) {
|
||||||
|
super.onPut(key, value)
|
||||||
|
logger.info { "$key with progress: $value." }
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}*/
|
||||||
@ -0,0 +1,54 @@
|
|||||||
|
package no.iktdev.streamit.content.encode
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
|
import no.iktdev.streamit.content.common.DefaultKafkaReader
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.DeserializerRegistry
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.EncodeWorkDeserializer
|
||||||
|
import no.iktdev.streamit.content.encode.runner.RunnerCoordinator
|
||||||
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
|
import no.iktdev.streamit.library.kafka.consumers.DefaultConsumer
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.SimpleMessageListener
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.deserializeIfSuccessful
|
||||||
|
import org.apache.kafka.clients.consumer.ConsumerRecord
|
||||||
|
import org.springframework.stereotype.Service
|
||||||
|
|
||||||
|
@Service
|
||||||
|
class EncodeWorkConsumer(private val runnerCoordinator: RunnerCoordinator) : DefaultKafkaReader("encodeWork") {
|
||||||
|
lateinit var encodeInstructionsListener: EncodeInformationListener
|
||||||
|
|
||||||
|
init {
|
||||||
|
encodeInstructionsListener = EncodeInformationListener(
|
||||||
|
topic = CommonConfig.kafkaTopic,
|
||||||
|
defaultConsumer,
|
||||||
|
accepts = listOf(KafkaEvents.EVENT_READER_ENCODE_GENERATED_VIDEO.event),
|
||||||
|
runnerCoordinator
|
||||||
|
)
|
||||||
|
encodeInstructionsListener.listen()
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun loadDeserializers(): Map<String, IMessageDataDeserialization<*>> {
|
||||||
|
return DeserializerRegistry.getEventToDeserializer(
|
||||||
|
KafkaEvents.EVENT_READER_ENCODE_GENERATED_VIDEO
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class EncodeInformationListener(
|
||||||
|
topic: String,
|
||||||
|
consumer: DefaultConsumer,
|
||||||
|
accepts: List<String>,
|
||||||
|
val runnerCoordinator: RunnerCoordinator
|
||||||
|
) : SimpleMessageListener(
|
||||||
|
topic, consumer,
|
||||||
|
accepts
|
||||||
|
) {
|
||||||
|
override fun onMessageReceived(data: ConsumerRecord<String, Message>) {
|
||||||
|
val message = data.value().apply {
|
||||||
|
this.data = EncodeWorkDeserializer().deserializeIfSuccessful(data.value())
|
||||||
|
}
|
||||||
|
runnerCoordinator.addEncodeMessageToQueue(message)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,55 @@
|
|||||||
|
package no.iktdev.streamit.content.encode
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
|
import no.iktdev.streamit.content.common.DefaultKafkaReader
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.DeserializerRegistry
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.ExtractWorkDeserializer
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.ExtractWork
|
||||||
|
import no.iktdev.streamit.content.encode.runner.RunnerCoordinator
|
||||||
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
|
import no.iktdev.streamit.library.kafka.consumers.DefaultConsumer
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.SimpleMessageListener
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.deserializeIfSuccessful
|
||||||
|
import org.apache.kafka.clients.consumer.ConsumerRecord
|
||||||
|
import org.springframework.stereotype.Service
|
||||||
|
|
||||||
|
@Service
|
||||||
|
class ExtractWorkConsumer(private val runnerCoordinator: RunnerCoordinator) : DefaultKafkaReader("extractWork") {
|
||||||
|
lateinit var encodeInstructionsListener: ExtractWorkListener
|
||||||
|
|
||||||
|
init {
|
||||||
|
encodeInstructionsListener = ExtractWorkListener(
|
||||||
|
topic = CommonConfig.kafkaTopic,
|
||||||
|
defaultConsumer,
|
||||||
|
accepts = listOf(KafkaEvents.EVENT_READER_ENCODE_GENERATED_SUBTITLE.event),
|
||||||
|
runnerCoordinator
|
||||||
|
)
|
||||||
|
encodeInstructionsListener.listen()
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun loadDeserializers(): Map<String, IMessageDataDeserialization<*>> {
|
||||||
|
return DeserializerRegistry.getEventToDeserializer(
|
||||||
|
KafkaEvents.EVENT_READER_ENCODE_GENERATED_SUBTITLE
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class ExtractWorkListener(
|
||||||
|
topic: String,
|
||||||
|
consumer: DefaultConsumer,
|
||||||
|
accepts: List<String>,
|
||||||
|
val runnerCoordinator: RunnerCoordinator
|
||||||
|
) : SimpleMessageListener(
|
||||||
|
topic, consumer,
|
||||||
|
accepts
|
||||||
|
) {
|
||||||
|
override fun onMessageReceived(data: ConsumerRecord<String, Message>) {
|
||||||
|
val message = data.value().apply {
|
||||||
|
this.data = ExtractWorkDeserializer().deserializeIfSuccessful(data.value())
|
||||||
|
}
|
||||||
|
runnerCoordinator.addExtractMessageToQueue(message)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,16 @@
|
|||||||
|
package no.iktdev.streamit.content.encode.progress
|
||||||
|
|
||||||
|
data class Progress(
|
||||||
|
val frame: Int?,
|
||||||
|
val fps: Double?,
|
||||||
|
val stream_0_0_q: Double?,
|
||||||
|
val bitrate: String?,
|
||||||
|
val total_size: Int?,
|
||||||
|
val out_time_us: Long?,
|
||||||
|
val out_time_ms: Long?,
|
||||||
|
val out_time: String?,
|
||||||
|
val dup_frames: Int?,
|
||||||
|
val drop_frames: Int?,
|
||||||
|
val speed: Double?,
|
||||||
|
val progress: String?
|
||||||
|
)
|
||||||
@ -0,0 +1,35 @@
|
|||||||
|
package no.iktdev.streamit.content.encode.progress
|
||||||
|
|
||||||
|
class ProgressDecoder {
|
||||||
|
fun parseVideoProgress(lines: List<String>): Progress? {
|
||||||
|
var frame: Int? = null
|
||||||
|
var progress: String? = null
|
||||||
|
val metadataMap = mutableMapOf<String, String>()
|
||||||
|
|
||||||
|
for (line in lines) {
|
||||||
|
val keyValuePairs = line.split(" ")
|
||||||
|
for (keyValuePair in keyValuePairs) {
|
||||||
|
val (key, value) = keyValuePair.split("=")
|
||||||
|
metadataMap[key] = value
|
||||||
|
}
|
||||||
|
|
||||||
|
if (frame == null) {
|
||||||
|
frame = metadataMap["frame"]?.toIntOrNull()
|
||||||
|
}
|
||||||
|
|
||||||
|
progress = metadataMap["progress"]
|
||||||
|
}
|
||||||
|
|
||||||
|
return if (progress != null) {
|
||||||
|
// When "progress" is found, build and return the VideoMetadata object
|
||||||
|
Progress(
|
||||||
|
frame, metadataMap["fps"]?.toDoubleOrNull(), metadataMap["stream_0_0_q"]?.toDoubleOrNull(),
|
||||||
|
metadataMap["bitrate"], metadataMap["total_size"]?.toIntOrNull(), metadataMap["out_time_us"]?.toLongOrNull(),
|
||||||
|
metadataMap["out_time_ms"]?.toLongOrNull(), metadataMap["out_time"], metadataMap["dup_frames"]?.toIntOrNull(),
|
||||||
|
metadataMap["drop_frames"]?.toIntOrNull(), metadataMap["speed"]?.toDoubleOrNull(), progress
|
||||||
|
)
|
||||||
|
} else {
|
||||||
|
null // If "progress" is not found, return null
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,68 @@
|
|||||||
|
package no.iktdev.streamit.content.encode.runner
|
||||||
|
|
||||||
|
import EncodeEnv
|
||||||
|
import no.iktdev.exfl.observable.ObservableList
|
||||||
|
import no.iktdev.exfl.observable.observableListOf
|
||||||
|
import no.iktdev.streamit.content.common.deamon.Daemon
|
||||||
|
import no.iktdev.streamit.content.common.deamon.IDaemon
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.EncodeWork
|
||||||
|
import no.iktdev.streamit.content.encode.progress.Progress
|
||||||
|
import no.iktdev.streamit.content.encode.progress.ProgressDecoder
|
||||||
|
import java.io.File
|
||||||
|
|
||||||
|
class EncodeDaemon(val referenceId: String, val work: EncodeWork, val daemonInterface: IEncodeListener): IDaemon {
|
||||||
|
var outputCache = observableListOf<String>()
|
||||||
|
private val decoder = ProgressDecoder()
|
||||||
|
private fun produceProgress(items: List<String>) {
|
||||||
|
val progress = decoder.parseVideoProgress(items)
|
||||||
|
if (progress != null) {
|
||||||
|
daemonInterface.onProgress(referenceId, work, progress)
|
||||||
|
outputCache.clear()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
init {
|
||||||
|
outputCache.addListener(object : ObservableList.Listener<String> {
|
||||||
|
override fun onAdded(item: String) {
|
||||||
|
produceProgress(outputCache)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
suspend fun runUsingWorkItem(): Int {
|
||||||
|
if (!File(work.outFile).parentFile.exists()) {
|
||||||
|
File(work.outFile).parentFile.mkdirs()
|
||||||
|
}
|
||||||
|
val adjustedArgs = listOf(
|
||||||
|
"-hide_banner", "-i", work.inFile, *work.arguments.toTypedArray(), work.outFile,
|
||||||
|
"-progress", "pipe:1"
|
||||||
|
) + if (EncodeEnv.allowOverwrite) listOf("-y") else emptyList()
|
||||||
|
return Daemon(EncodeEnv.ffmpeg, this).run(adjustedArgs)
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onStarted() {
|
||||||
|
super.onStarted()
|
||||||
|
daemonInterface.onStarted(referenceId, work)
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onEnded() {
|
||||||
|
super.onEnded()
|
||||||
|
daemonInterface.onEnded(referenceId, work)
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onError(code: Int) {
|
||||||
|
daemonInterface.onError(referenceId, work, code)
|
||||||
|
}
|
||||||
|
override fun onOutputChanged(line: String) {
|
||||||
|
super.onOutputChanged(line)
|
||||||
|
outputCache.add(line)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
interface IEncodeListener {
|
||||||
|
fun onStarted(referenceId: String, work: EncodeWork)
|
||||||
|
fun onError(referenceId: String, work: EncodeWork, code: Int)
|
||||||
|
fun onProgress(referenceId: String, work: EncodeWork, progress: Progress)
|
||||||
|
fun onEnded(referenceId: String, work: EncodeWork)
|
||||||
|
}
|
||||||
@ -0,0 +1,54 @@
|
|||||||
|
package no.iktdev.streamit.content.encode.runner
|
||||||
|
|
||||||
|
import EncodeEnv
|
||||||
|
import no.iktdev.exfl.observable.ObservableList
|
||||||
|
import no.iktdev.exfl.observable.observableListOf
|
||||||
|
import no.iktdev.streamit.content.common.deamon.Daemon
|
||||||
|
import no.iktdev.streamit.content.common.deamon.IDaemon
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.EncodeWork
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.ExtractWork
|
||||||
|
import no.iktdev.streamit.content.encode.progress.Progress
|
||||||
|
import no.iktdev.streamit.content.encode.progress.ProgressDecoder
|
||||||
|
import java.io.File
|
||||||
|
|
||||||
|
class ExtractDaemon(val referenceId: String, val work: ExtractWork, val daemonInterface: IExtractListener): IDaemon {
|
||||||
|
var outputCache = observableListOf<String>()
|
||||||
|
|
||||||
|
|
||||||
|
suspend fun runUsingWorkItem(): Int {
|
||||||
|
if (!File(work.outFile).parentFile.exists()) {
|
||||||
|
File(work.outFile).parentFile.mkdirs()
|
||||||
|
}
|
||||||
|
val adjustedArgs = listOf(
|
||||||
|
"-hide_banner", "-i", work.inFile, *work.arguments.toTypedArray(), work.outFile,
|
||||||
|
"-progress", "pipe:1"
|
||||||
|
) + if (EncodeEnv.allowOverwrite) listOf("-y") else emptyList()
|
||||||
|
return Daemon(EncodeEnv.ffmpeg, this).run(adjustedArgs)
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onStarted() {
|
||||||
|
super.onStarted()
|
||||||
|
daemonInterface.onStarted(referenceId, work)
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onEnded() {
|
||||||
|
super.onEnded()
|
||||||
|
daemonInterface.onEnded(referenceId, work)
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onError(code: Int) {
|
||||||
|
daemonInterface.onError(referenceId, work, code)
|
||||||
|
}
|
||||||
|
override fun onOutputChanged(line: String) {
|
||||||
|
super.onOutputChanged(line)
|
||||||
|
outputCache.add(line)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
interface IExtractListener {
|
||||||
|
fun onStarted(referenceId: String, work: ExtractWork)
|
||||||
|
fun onError(referenceId: String, work: ExtractWork, code: Int)
|
||||||
|
fun onProgress(referenceId: String, work: ExtractWork, progress: Progress) {}
|
||||||
|
fun onEnded(referenceId: String, work: ExtractWork)
|
||||||
|
}
|
||||||
@ -0,0 +1,109 @@
|
|||||||
|
package no.iktdev.streamit.content.encode.runner
|
||||||
|
|
||||||
|
import EncodeEnv
|
||||||
|
import kotlinx.coroutines.runBlocking
|
||||||
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
|
import no.iktdev.streamit.content.common.deamon.IDaemon
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.EncodeWork
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.ExtractWork
|
||||||
|
import no.iktdev.streamit.content.encode.progress.Progress
|
||||||
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Status
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.StatusType
|
||||||
|
import no.iktdev.streamit.library.kafka.producer.DefaultProducer
|
||||||
|
import org.springframework.stereotype.Service
|
||||||
|
import java.util.concurrent.ExecutorService
|
||||||
|
import java.util.concurrent.LinkedBlockingQueue
|
||||||
|
import java.util.concurrent.ThreadPoolExecutor
|
||||||
|
import java.util.concurrent.TimeUnit
|
||||||
|
|
||||||
|
@Service
|
||||||
|
class RunnerCoordinator {
|
||||||
|
|
||||||
|
val producer = DefaultProducer(CommonConfig.kafkaTopic)
|
||||||
|
|
||||||
|
val encodeExecutor: ExecutorService = ThreadPoolExecutor(
|
||||||
|
EncodeEnv.maxRunners,
|
||||||
|
EncodeEnv.maxRunners,
|
||||||
|
0L,
|
||||||
|
TimeUnit.MILLISECONDS,
|
||||||
|
LinkedBlockingQueue()
|
||||||
|
)
|
||||||
|
|
||||||
|
val extractExecutor: ExecutorService = ThreadPoolExecutor(
|
||||||
|
EncodeEnv.maxRunners,
|
||||||
|
EncodeEnv.maxRunners,
|
||||||
|
0L,
|
||||||
|
TimeUnit.MILLISECONDS,
|
||||||
|
LinkedBlockingQueue()
|
||||||
|
)
|
||||||
|
|
||||||
|
fun addEncodeMessageToQueue(message: Message) {
|
||||||
|
encodeExecutor.execute {
|
||||||
|
runBlocking {
|
||||||
|
if (message.data is EncodeWork) {
|
||||||
|
val data: EncodeWork = message.data as EncodeWork
|
||||||
|
val encodeDaemon = EncodeDaemon(message.referenceId, data, encodeListener)
|
||||||
|
encodeDaemon.runUsingWorkItem()
|
||||||
|
} else {
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_STARTED_VIDEO_FILE.event, message.withNewStatus(Status(StatusType.ERROR, "Data is not an instance of EncodeWork")))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_STARTED_VIDEO_FILE.event, message.withNewStatus(Status(StatusType.PENDING)))
|
||||||
|
}
|
||||||
|
|
||||||
|
fun addExtractMessageToQueue(message: Message) {
|
||||||
|
extractExecutor.execute {
|
||||||
|
runBlocking {
|
||||||
|
if (message.data is ExtractWork) {
|
||||||
|
val data: ExtractWork = message.data as ExtractWork
|
||||||
|
val extractDaemon = ExtractDaemon(message.referenceId, data, extractListener)
|
||||||
|
extractDaemon.runUsingWorkItem()
|
||||||
|
} else {
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_STARTED_SUBTITLE_FILE.event, message.withNewStatus(Status(StatusType.ERROR, "Data is not an instance of ExtractWork")))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_STARTED_SUBTITLE_FILE.event, message.withNewStatus(Status(StatusType.PENDING)))
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
val encodeListener = object: IEncodeListener {
|
||||||
|
override fun onStarted(referenceId: String, work: EncodeWork) {
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_STARTED_VIDEO_FILE.event, Message(referenceId, Status(StatusType.SUCCESS), work))
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onError(referenceId: String, work: EncodeWork, code: Int) {
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_ENDED_VIDEO_FILE.event, Message(referenceId, Status(StatusType.ERROR, message = code.toString()), work))
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onProgress(referenceId: String, work: EncodeWork, progress: Progress) {
|
||||||
|
// TODO: Pass to listener
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onEnded(referenceId: String, work: EncodeWork) {
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_ENDED_VIDEO_FILE.event, Message(referenceId, Status(StatusType.SUCCESS), work))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
val extractListener = object : IExtractListener {
|
||||||
|
override fun onStarted(referenceId: String, work: ExtractWork) {
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_STARTED_SUBTITLE_FILE.event, Message(referenceId, Status(StatusType.SUCCESS), work))
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onError(referenceId: String, work: ExtractWork, code: Int) {
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_ENDED_SUBTITLE_FILE.event, Message(referenceId, Status(StatusType.ERROR), work))
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onEnded(referenceId: String, work: ExtractWork) {
|
||||||
|
producer.sendMessage(KafkaEvents.EVENT_ENCODER_ENDED_SUBTITLE_FILE.event, Message(referenceId, Status(StatusType.SUCCESS), work))
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
@ -0,0 +1,33 @@
|
|||||||
|
package no.iktdev.streamit.content.encode.progress
|
||||||
|
|
||||||
|
class ProgressDecoderTest {
|
||||||
|
|
||||||
|
|
||||||
|
val text = """
|
||||||
|
frame=16811 fps= 88 q=40.0 size= 9984kB time=00:01:10.79 bitrate=1155.3kbits/s speed=3.71x
|
||||||
|
fps=88.03
|
||||||
|
stream_0_0_q=40.0
|
||||||
|
bitrate=1155.3kbits/s
|
||||||
|
total_size=10223752
|
||||||
|
out_time_us=70798005
|
||||||
|
out_time_ms=70798005
|
||||||
|
out_time=00:01:10.798005
|
||||||
|
dup_frames=0
|
||||||
|
drop_frames=0
|
||||||
|
speed=3.71x
|
||||||
|
progress=continue
|
||||||
|
frame= 1710 fps= 84 q=-1.0 Lsize= 12124kB time=00:01:11.91 bitrate=1381.2kbits/s speed=3.53x
|
||||||
|
frame=1710
|
||||||
|
fps=84.01
|
||||||
|
stream_0_0_q=-1.0
|
||||||
|
bitrate=1381.2kbits/s
|
||||||
|
total_size=12415473
|
||||||
|
out_time_us=71910998
|
||||||
|
out_time_ms=71910998
|
||||||
|
out_time=00:01:11.910998
|
||||||
|
dup_frames=0
|
||||||
|
drop_frames=0
|
||||||
|
speed=3.53x
|
||||||
|
progress=end
|
||||||
|
""".trimIndent()
|
||||||
|
}
|
||||||
@ -1,8 +1,4 @@
|
|||||||
FROM bskjon/azuljava:17
|
FROM bskjon/debian-azuljava17-ffmpeg:latest
|
||||||
EXPOSE 8080
|
EXPOSE 8080
|
||||||
|
|
||||||
RUN mkdir -p /src/input
|
|
||||||
RUN apt update -y
|
|
||||||
RUN apt install -y ffmpeg
|
|
||||||
|
|
||||||
COPY ./build/libs/reader.jar /usr/share/app/app.jar
|
COPY ./build/libs/reader.jar /usr/share/app/app.jar
|
||||||
@ -23,8 +23,11 @@ repositories {
|
|||||||
}
|
}
|
||||||
|
|
||||||
dependencies {
|
dependencies {
|
||||||
implementation("no.iktdev.streamit.library:streamit-library-kafka:0.0.2-alpha48")
|
implementation("no.iktdev.streamit.library:streamit-library-kafka:0.0.2-alpha63")
|
||||||
implementation("no.iktdev:exfl:0.0.8-SNAPSHOT")
|
implementation("no.iktdev:exfl:0.0.12-SNAPSHOT")
|
||||||
|
|
||||||
|
implementation("no.iktdev.streamit.library:streamit-library-db:0.0.6-alpha7")
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
implementation("com.github.pgreze:kotlin-process:1.3.1")
|
implementation("com.github.pgreze:kotlin-process:1.3.1")
|
||||||
|
|||||||
@ -2,14 +2,20 @@ package no.iktdev.streamit.content.reader.analyzer.contentDeterminator
|
|||||||
|
|
||||||
import mu.KotlinLogging
|
import mu.KotlinLogging
|
||||||
import no.iktdev.streamit.content.common.CommonConfig
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
|
import no.iktdev.streamit.content.common.DefaultKafkaReader
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.FileResultDeserializer
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.MetadataResultDeserializer
|
||||||
import no.iktdev.streamit.content.common.dto.ContentOutName
|
import no.iktdev.streamit.content.common.dto.ContentOutName
|
||||||
import no.iktdev.streamit.content.common.dto.Metadata
|
import no.iktdev.streamit.content.common.dto.Metadata
|
||||||
import no.iktdev.streamit.content.reader.DefaultKafkaReader
|
import no.iktdev.streamit.content.common.dto.reader.EpisodeInfo
|
||||||
import no.iktdev.streamit.content.reader.fileWatcher.FileWatcher
|
import no.iktdev.streamit.content.common.dto.reader.FileResult
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.MovieInfo
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.VideoInfo
|
||||||
import no.iktdev.streamit.library.kafka.KafkaEvents
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
import no.iktdev.streamit.library.kafka.dto.Message
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
import no.iktdev.streamit.library.kafka.dto.Status
|
import no.iktdev.streamit.library.kafka.dto.Status
|
||||||
import no.iktdev.streamit.library.kafka.dto.StatusType
|
import no.iktdev.streamit.library.kafka.dto.StatusType
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
import no.iktdev.streamit.library.kafka.listener.sequential.ISequentialMessageEvent
|
import no.iktdev.streamit.library.kafka.listener.sequential.ISequentialMessageEvent
|
||||||
import no.iktdev.streamit.library.kafka.listener.sequential.SequentialMessageListener
|
import no.iktdev.streamit.library.kafka.listener.sequential.SequentialMessageListener
|
||||||
import org.springframework.stereotype.Service
|
import org.springframework.stereotype.Service
|
||||||
@ -24,8 +30,8 @@ class ContentDeterminate: DefaultKafkaReader("contentDeterminate"), ISequentialM
|
|||||||
consumer = defaultConsumer,
|
consumer = defaultConsumer,
|
||||||
accept = KafkaEvents.EVENT_READER_RECEIVED_FILE.event,
|
accept = KafkaEvents.EVENT_READER_RECEIVED_FILE.event,
|
||||||
subAccepts = listOf(KafkaEvents.EVENT_METADATA_OBTAINED.event),
|
subAccepts = listOf(KafkaEvents.EVENT_METADATA_OBTAINED.event),
|
||||||
deserializers = Deserializers().getDeserializers(),
|
deserializers = loadDeserializers(),
|
||||||
this
|
listener = this
|
||||||
) {}
|
) {}
|
||||||
|
|
||||||
init {
|
init {
|
||||||
@ -42,20 +48,20 @@ class ContentDeterminate: DefaultKafkaReader("contentDeterminate"), ISequentialM
|
|||||||
logger.info { "All messages are received" }
|
logger.info { "All messages are received" }
|
||||||
|
|
||||||
val initMessage = result[KafkaEvents.EVENT_READER_RECEIVED_FILE.event]
|
val initMessage = result[KafkaEvents.EVENT_READER_RECEIVED_FILE.event]
|
||||||
if (initMessage == null) {
|
if (initMessage == null || initMessage.status.statusType != StatusType.SUCCESS) {
|
||||||
produceErrorMessage(Message(referenceId = referenceId, status = Status(statusType = StatusType.ERROR)), "Initiator message not found!")
|
produceErrorMessage(KafkaEvents.EVENT_READER_DETERMINED_FILENAME, Message(referenceId = referenceId, status = Status(statusType = StatusType.ERROR)), "Initiator message not found!")
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
val fileResult = initMessage.data as FileWatcher.FileResult?
|
val fileResult = initMessage.data as FileResult?
|
||||||
if (fileResult == null) {
|
if (fileResult == null) {
|
||||||
produceErrorMessage(initMessage, "FileResult is either null or not deserializable!")
|
produceErrorMessage(KafkaEvents.EVENT_READER_DETERMINED_FILENAME, initMessage, "FileResult is either null or not deserializable!")
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
val metadataMessage = result[KafkaEvents.EVENT_METADATA_OBTAINED.event]
|
val metadataMessage = result[KafkaEvents.EVENT_METADATA_OBTAINED.event]
|
||||||
val metadata = if (metadataMessage?.status?.statusType == StatusType.SUCCESS) metadataMessage.data as Metadata? else null
|
val metadata = if (metadataMessage?.status?.statusType == StatusType.SUCCESS) metadataMessage.data as Metadata? else null
|
||||||
|
|
||||||
val baseFileName = if (metadata?.type == null) {
|
val videoInfo = if (metadata?.type == null) {
|
||||||
FileNameDeterminate(fileResult.title, fileResult.sanitizedName).getDeterminedFileName()
|
FileNameDeterminate(fileResult.title, fileResult.sanitizedName).getDeterminedFileName()
|
||||||
} else if (metadata.type.lowercase() == "movie") {
|
} else if (metadata.type.lowercase() == "movie") {
|
||||||
FileNameDeterminate(fileResult.title, fileResult.sanitizedName, FileNameDeterminate.ContentType.MOVIE).getDeterminedFileName()
|
FileNameDeterminate(fileResult.title, fileResult.sanitizedName, FileNameDeterminate.ContentType.MOVIE).getDeterminedFileName()
|
||||||
@ -63,9 +69,26 @@ class ContentDeterminate: DefaultKafkaReader("contentDeterminate"), ISequentialM
|
|||||||
FileNameDeterminate(fileResult.title, fileResult.sanitizedName, FileNameDeterminate.ContentType.SERIE).getDeterminedFileName()
|
FileNameDeterminate(fileResult.title, fileResult.sanitizedName, FileNameDeterminate.ContentType.SERIE).getDeterminedFileName()
|
||||||
}
|
}
|
||||||
|
|
||||||
val out = ContentOutName(baseFileName)
|
if (videoInfo == null) {
|
||||||
|
produceErrorMessage(KafkaEvents.EVENT_READER_DETERMINED_FILENAME, initMessage, "VideoInfo is null." )
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
val out = ContentOutName(videoInfo.fullName)
|
||||||
produceMessage(KafkaEvents.EVENT_READER_DETERMINED_FILENAME, initMessage, out)
|
produceMessage(KafkaEvents.EVENT_READER_DETERMINED_FILENAME, initMessage, out)
|
||||||
|
|
||||||
|
if (videoInfo is EpisodeInfo) {
|
||||||
|
produceMessage(KafkaEvents.EVENT_READER_DETERMINED_SERIE, initMessage, videoInfo)
|
||||||
|
} else if (videoInfo is MovieInfo) {
|
||||||
|
produceMessage(KafkaEvents.EVENT_READER_DETERMINED_MOVIE, initMessage, videoInfo)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
final override fun loadDeserializers(): Map<String, IMessageDataDeserialization<*>> {
|
||||||
|
return mutableMapOf(
|
||||||
|
KafkaEvents.EVENT_READER_RECEIVED_FILE.event to FileResultDeserializer(),
|
||||||
|
KafkaEvents.EVENT_METADATA_OBTAINED.event to MetadataResultDeserializer()
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
@ -1,37 +0,0 @@
|
|||||||
package no.iktdev.streamit.content.reader.analyzer.contentDeterminator
|
|
||||||
|
|
||||||
import no.iktdev.streamit.content.common.dto.Metadata
|
|
||||||
import no.iktdev.streamit.content.reader.fileWatcher.FileWatcher
|
|
||||||
import no.iktdev.streamit.library.kafka.KafkaEvents
|
|
||||||
import no.iktdev.streamit.library.kafka.dto.Message
|
|
||||||
import no.iktdev.streamit.library.kafka.dto.StatusType
|
|
||||||
import no.iktdev.streamit.library.kafka.listener.sequential.IMessageDataDeserialization
|
|
||||||
|
|
||||||
class Deserializers {
|
|
||||||
|
|
||||||
val fileReceived = object : IMessageDataDeserialization<FileWatcher.FileResult> {
|
|
||||||
override fun deserialize(incomingMessage: Message): FileWatcher.FileResult? {
|
|
||||||
if (incomingMessage.status.statusType != StatusType.SUCCESS) {
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
return incomingMessage.dataAs(FileWatcher.FileResult::class.java)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
val metadataReceived = object: IMessageDataDeserialization<Metadata> {
|
|
||||||
override fun deserialize(incomingMessage: Message): Metadata? {
|
|
||||||
if (incomingMessage.status.statusType != StatusType.SUCCESS) {
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
return incomingMessage.dataAs(Metadata::class.java)
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
fun getDeserializers(): Map<String, IMessageDataDeserialization<*>> {
|
|
||||||
return mutableMapOf(
|
|
||||||
KafkaEvents.EVENT_READER_RECEIVED_FILE.event to fileReceived,
|
|
||||||
KafkaEvents.EVENT_METADATA_OBTAINED.event to metadataReceived
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
@ -1,5 +1,9 @@
|
|||||||
package no.iktdev.streamit.content.reader.analyzer.contentDeterminator
|
package no.iktdev.streamit.content.reader.analyzer.contentDeterminator
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.EpisodeInfo
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.MovieInfo
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.VideoInfo
|
||||||
|
|
||||||
class FileNameDeterminate(val title: String, val sanitizedName: String, val ctype: ContentType = ContentType.UNDEFINED) {
|
class FileNameDeterminate(val title: String, val sanitizedName: String, val ctype: ContentType = ContentType.UNDEFINED) {
|
||||||
|
|
||||||
enum class ContentType {
|
enum class ContentType {
|
||||||
@ -8,7 +12,7 @@ class FileNameDeterminate(val title: String, val sanitizedName: String, val ctyp
|
|||||||
UNDEFINED
|
UNDEFINED
|
||||||
}
|
}
|
||||||
|
|
||||||
fun getDeterminedFileName(): String {
|
fun getDeterminedFileName(): VideoInfo? {
|
||||||
return when (ctype) {
|
return when (ctype) {
|
||||||
ContentType.MOVIE -> determineMovieFileName()
|
ContentType.MOVIE -> determineMovieFileName()
|
||||||
ContentType.SERIE -> determineSerieFileName()
|
ContentType.SERIE -> determineSerieFileName()
|
||||||
@ -16,23 +20,23 @@ class FileNameDeterminate(val title: String, val sanitizedName: String, val ctyp
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private fun determineMovieFileName(): String {
|
private fun determineMovieFileName(): MovieInfo? {
|
||||||
val movieEx = MovieEx(title, sanitizedName)
|
val movieEx = MovieEx(title, sanitizedName)
|
||||||
val result = when {
|
val result = when {
|
||||||
movieEx.isDefinedWithYear() != null -> sanitizedName.replace(movieEx.isDefinedWithYear()!!, "").trim()
|
movieEx.isDefinedWithYear() != null -> sanitizedName.replace(movieEx.isDefinedWithYear()!!, "").trim()
|
||||||
movieEx.doesContainMovieKeywords() -> sanitizedName.replace(Regex("(?i)\\s*\\(\\s*movie\\s*\\)\\s*"), "").trim()
|
movieEx.doesContainMovieKeywords() -> sanitizedName.replace(Regex("(?i)\\s*\\(\\s*movie\\s*\\)\\s*"), "").trim()
|
||||||
else -> title
|
else -> title
|
||||||
}
|
}
|
||||||
return result
|
return MovieInfo(title, result)
|
||||||
}
|
}
|
||||||
|
|
||||||
private fun determineSerieFileName(): String {
|
private fun determineSerieFileName(): EpisodeInfo? {
|
||||||
val serieEx = SerieEx(title, sanitizedName)
|
val serieEx = SerieEx(title, sanitizedName)
|
||||||
val (season, episode) = serieEx.findSeasonAndEpisode(sanitizedName)
|
val (season, episode) = serieEx.findSeasonAndEpisode(sanitizedName)
|
||||||
val episodeNumberSingle = serieEx.findEpisodeNumber()
|
val episodeNumberSingle = serieEx.findEpisodeNumber()
|
||||||
|
|
||||||
val seasonNumber = season ?: "1"
|
val seasonNumber = season ?: "1"
|
||||||
val episodeNumber = episode ?: (episodeNumberSingle ?: return sanitizedName)
|
val episodeNumber = episode ?: (episodeNumberSingle ?: return null)
|
||||||
val seasonEpisodeCombined = serieEx.getSeasonEpisodeCombined(seasonNumber, episodeNumber)
|
val seasonEpisodeCombined = serieEx.getSeasonEpisodeCombined(seasonNumber, episodeNumber)
|
||||||
val episodeTitle = serieEx.findEpisodeTitle()
|
val episodeTitle = serieEx.findEpisodeTitle()
|
||||||
|
|
||||||
@ -50,11 +54,11 @@ class FileNameDeterminate(val title: String, val sanitizedName: String, val ctyp
|
|||||||
|
|
||||||
}
|
}
|
||||||
} else title
|
} else title
|
||||||
|
val fullName = "${useTitle.trim()} - $seasonEpisodeCombined ${if (episodeTitle.isNullOrEmpty()) "" else "- $episodeTitle"}".trim()
|
||||||
return "${useTitle.trim()} - $seasonEpisodeCombined ${if (episodeTitle.isNullOrEmpty()) "" else "- $episodeTitle"}".trim()
|
return EpisodeInfo(title, episodeNumber.toInt(), seasonNumber.toInt(), episodeTitle, fullName)
|
||||||
}
|
}
|
||||||
|
|
||||||
private fun determineUndefinedFileName(): String {
|
private fun determineUndefinedFileName(): VideoInfo? {
|
||||||
val serieEx = SerieEx(title, sanitizedName)
|
val serieEx = SerieEx(title, sanitizedName)
|
||||||
val (season, episode) = serieEx.findSeasonAndEpisode(sanitizedName)
|
val (season, episode) = serieEx.findSeasonAndEpisode(sanitizedName)
|
||||||
return if (sanitizedName.contains(" - ") || season != null || episode != null) {
|
return if (sanitizedName.contains(" - ") || season != null || episode != null) {
|
||||||
|
|||||||
@ -1,86 +0,0 @@
|
|||||||
package no.iktdev.streamit.content.reader.analyzer.encoding
|
|
||||||
|
|
||||||
import com.google.gson.Gson
|
|
||||||
import com.google.gson.JsonObject
|
|
||||||
import no.iktdev.streamit.content.common.dto.ContentOutName
|
|
||||||
import no.iktdev.streamit.content.common.streams.AudioStream
|
|
||||||
import no.iktdev.streamit.content.common.streams.MediaStreams
|
|
||||||
import no.iktdev.streamit.content.common.streams.SubtitleStream
|
|
||||||
import no.iktdev.streamit.content.common.streams.VideoStream
|
|
||||||
import no.iktdev.streamit.content.reader.fileWatcher.FileWatcher
|
|
||||||
import no.iktdev.streamit.library.kafka.KafkaEvents
|
|
||||||
import no.iktdev.streamit.library.kafka.dto.Message
|
|
||||||
import no.iktdev.streamit.library.kafka.dto.StatusType
|
|
||||||
import no.iktdev.streamit.library.kafka.listener.sequential.IMessageDataDeserialization
|
|
||||||
|
|
||||||
class EncodedDeserializers {
|
|
||||||
val gson = Gson()
|
|
||||||
|
|
||||||
val fileReceived = object : IMessageDataDeserialization<FileWatcher.FileResult> {
|
|
||||||
override fun deserialize(incomingMessage: Message): FileWatcher.FileResult? {
|
|
||||||
if (incomingMessage.status.statusType != StatusType.SUCCESS) {
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
return incomingMessage.dataAs(FileWatcher.FileResult::class.java)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
val determinedFileNameReceived = object: IMessageDataDeserialization<ContentOutName> {
|
|
||||||
override fun deserialize(incomingMessage: Message): ContentOutName? {
|
|
||||||
if (incomingMessage.status.statusType != StatusType.SUCCESS) {
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
return incomingMessage.dataAs(ContentOutName::class.java)
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
val mediaStreams = object : IMessageDataDeserialization<MediaStreams> {
|
|
||||||
override fun deserialize(incomingMessage: Message): MediaStreams? {
|
|
||||||
return try {
|
|
||||||
if (incomingMessage.status.statusType != StatusType.SUCCESS) {
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
val jsonObject = if (incomingMessage.data is String) {
|
|
||||||
gson.fromJson(incomingMessage.data as String, JsonObject::class.java)
|
|
||||||
} else {
|
|
||||||
gson.fromJson(incomingMessage.dataAsJson(), JsonObject::class.java)
|
|
||||||
}
|
|
||||||
|
|
||||||
val streamsJsonArray = jsonObject.getAsJsonArray("streams")
|
|
||||||
|
|
||||||
val rstreams = streamsJsonArray.mapNotNull { streamJson ->
|
|
||||||
val streamObject = streamJson.asJsonObject
|
|
||||||
|
|
||||||
val codecType = streamObject.get("codec_type").asString
|
|
||||||
if (streamObject.has("codec_name") && streamObject.get("codec_name").asString == "mjpeg") {
|
|
||||||
null
|
|
||||||
} else {
|
|
||||||
when (codecType) {
|
|
||||||
"video" -> gson.fromJson(streamObject, VideoStream::class.java)
|
|
||||||
"audio" -> gson.fromJson(streamObject, AudioStream::class.java)
|
|
||||||
"subtitle" -> gson.fromJson(streamObject, SubtitleStream::class.java)
|
|
||||||
else -> null //throw IllegalArgumentException("Unknown stream type: $codecType")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return MediaStreams(rstreams)
|
|
||||||
} catch (e: Exception) {
|
|
||||||
e.printStackTrace()
|
|
||||||
null
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
fun getDeserializers(): Map<String, IMessageDataDeserialization<*>> {
|
|
||||||
return mutableMapOf(
|
|
||||||
KafkaEvents.EVENT_READER_RECEIVED_FILE.event to fileReceived,
|
|
||||||
KafkaEvents.EVENT_READER_RECEIVED_STREAMS.event to mediaStreams,
|
|
||||||
KafkaEvents.EVENT_READER_DETERMINED_FILENAME.event to determinedFileNameReceived
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
@ -2,30 +2,27 @@ package no.iktdev.streamit.content.reader.analyzer.encoding
|
|||||||
|
|
||||||
import mu.KotlinLogging
|
import mu.KotlinLogging
|
||||||
import no.iktdev.streamit.content.common.CommonConfig
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
|
import no.iktdev.streamit.content.common.DefaultKafkaReader
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.DeserializerRegistry
|
||||||
import no.iktdev.streamit.content.common.dto.ContentOutName
|
import no.iktdev.streamit.content.common.dto.ContentOutName
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.FileResult
|
||||||
import no.iktdev.streamit.content.common.streams.MediaStreams
|
import no.iktdev.streamit.content.common.streams.MediaStreams
|
||||||
import no.iktdev.streamit.content.reader.analyzer.encoding.dto.EncodeInformation
|
|
||||||
import no.iktdev.streamit.content.reader.analyzer.encoding.helpers.EncodeArgumentSelector
|
import no.iktdev.streamit.content.reader.analyzer.encoding.helpers.EncodeArgumentSelector
|
||||||
import no.iktdev.streamit.content.reader.fileWatcher.FileWatcher
|
|
||||||
import no.iktdev.streamit.library.kafka.KafkaEvents
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
import no.iktdev.streamit.library.kafka.consumers.DefaultConsumer
|
|
||||||
import no.iktdev.streamit.library.kafka.dto.Message
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
import no.iktdev.streamit.library.kafka.dto.Status
|
import no.iktdev.streamit.library.kafka.dto.Status
|
||||||
import no.iktdev.streamit.library.kafka.dto.StatusType
|
import no.iktdev.streamit.library.kafka.dto.StatusType
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
import no.iktdev.streamit.library.kafka.listener.sequential.ISequentialMessageEvent
|
import no.iktdev.streamit.library.kafka.listener.sequential.ISequentialMessageEvent
|
||||||
import no.iktdev.streamit.library.kafka.listener.sequential.SequentialMessageListener
|
import no.iktdev.streamit.library.kafka.listener.sequential.SequentialMessageListener
|
||||||
import no.iktdev.streamit.library.kafka.producer.DefaultProducer
|
|
||||||
import org.springframework.stereotype.Service
|
import org.springframework.stereotype.Service
|
||||||
import java.io.File
|
import java.io.File
|
||||||
|
|
||||||
private val logger = KotlinLogging.logger {}
|
private val logger = KotlinLogging.logger {}
|
||||||
|
|
||||||
@Service
|
@Service
|
||||||
class EncodedStreams : ISequentialMessageEvent {
|
class EncodedStreams : DefaultKafkaReader("encodedStreams"), ISequentialMessageEvent {
|
||||||
|
|
||||||
val messageProducer = DefaultProducer(CommonConfig.kafkaTopic)
|
|
||||||
|
|
||||||
final val defaultConsumer = DefaultConsumer(subId = "encodedStreams")
|
|
||||||
|
|
||||||
|
|
||||||
final val mainListener = object : SequentialMessageListener(
|
final val mainListener = object : SequentialMessageListener(
|
||||||
@ -36,8 +33,8 @@ class EncodedStreams : ISequentialMessageEvent {
|
|||||||
KafkaEvents.EVENT_READER_RECEIVED_STREAMS.event,
|
KafkaEvents.EVENT_READER_RECEIVED_STREAMS.event,
|
||||||
KafkaEvents.EVENT_READER_DETERMINED_FILENAME.event
|
KafkaEvents.EVENT_READER_DETERMINED_FILENAME.event
|
||||||
),
|
),
|
||||||
deserializers = EncodedDeserializers().getDeserializers(),
|
deserializers = loadDeserializers(),
|
||||||
this
|
listener = this
|
||||||
) {}
|
) {}
|
||||||
|
|
||||||
init {
|
init {
|
||||||
@ -51,69 +48,87 @@ class EncodedStreams : ISequentialMessageEvent {
|
|||||||
|
|
||||||
override fun onAllMessagesProcessed(referenceId: String, result: Map<String, Message?>) {
|
override fun onAllMessagesProcessed(referenceId: String, result: Map<String, Message?>) {
|
||||||
logger.info { "All messages are received" }
|
logger.info { "All messages are received" }
|
||||||
val baseMessage = result[KafkaEvents.EVENT_READER_RECEIVED_FILE.event]
|
val fileResultEvent = result[KafkaEvents.EVENT_READER_RECEIVED_FILE.event]
|
||||||
if (baseMessage == null) {
|
val determinedFileNameEvent = result[KafkaEvents.EVENT_READER_DETERMINED_FILENAME.event]
|
||||||
produceErrorMessage(
|
val streamEvent = result[KafkaEvents.EVENT_READER_RECEIVED_STREAMS.event]
|
||||||
Message(referenceId = referenceId, status = Status(statusType = StatusType.ERROR)),
|
|
||||||
"Initiator message not found!"
|
val fileResult = if (fileResultEvent != null && fileResultEvent.isSuccessful()) {
|
||||||
)
|
fileResultEvent.data as FileResult?
|
||||||
return
|
} else null
|
||||||
|
|
||||||
|
val outFileNameWithoutExtension = if (determinedFileNameEvent != null && determinedFileNameEvent.isSuccessful()) {
|
||||||
|
(determinedFileNameEvent.data as ContentOutName).baseName
|
||||||
|
} else fileResult?.sanitizedName
|
||||||
|
|
||||||
|
val streams = if (streamEvent != null && streamEvent.isSuccessful()) {
|
||||||
|
streamEvent.data as MediaStreams
|
||||||
|
} else null
|
||||||
|
|
||||||
|
createEncodeWork(referenceId, fileResult?.title, fileResult?.file, streams, outFileNameWithoutExtension)
|
||||||
|
createExtractWork(referenceId, fileResult?.title, fileResult?.file, streams, outFileNameWithoutExtension)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (result.values.any { it?.status?.statusType != StatusType.SUCCESS }) {
|
fun createEncodeWork(referenceId: String, collection: String?, inFile: String?, streams: MediaStreams?, outFileName: String?) {
|
||||||
produceErrorMessage(
|
if (inFile.isNullOrBlank()) {
|
||||||
Message(referenceId = referenceId, status = Status(statusType = StatusType.ERROR)),
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_VIDEO, referenceId, "No input file received"); return
|
||||||
"Failed messages found!"
|
|
||||||
)
|
|
||||||
return
|
|
||||||
}
|
}
|
||||||
val fileResult = baseMessage.data as FileWatcher.FileResult?
|
|
||||||
if (fileResult == null) {
|
|
||||||
produceErrorMessage(baseMessage, "FileResult is either null or not deserializable!")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
val determinedfnm = result[KafkaEvents.EVENT_READER_DETERMINED_FILENAME.event]
|
|
||||||
val determinedFileName = determinedfnm?.data as ContentOutName
|
|
||||||
|
|
||||||
val outFileName = if (determinedfnm.status.statusType == StatusType.SUCCESS)
|
|
||||||
determinedFileName.baseName
|
|
||||||
else fileResult.sanitizedName.ifBlank { File(fileResult.file).nameWithoutExtension }
|
|
||||||
|
|
||||||
|
|
||||||
val streams = result[KafkaEvents.EVENT_READER_RECEIVED_STREAMS.event]?.data as MediaStreams?
|
|
||||||
if (streams == null) {
|
if (streams == null) {
|
||||||
produceErrorMessage(baseMessage, "No streams received!")
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_VIDEO, referenceId, "No input streams received"); return
|
||||||
return
|
}
|
||||||
|
if (outFileName.isNullOrBlank()) {
|
||||||
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_VIDEO, referenceId, "No output file name received!"); return
|
||||||
|
}
|
||||||
|
if (collection.isNullOrBlank()) {
|
||||||
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_VIDEO, referenceId, "No collection provided for file!"); return
|
||||||
}
|
}
|
||||||
|
|
||||||
val encodeInformation =
|
val encodeInformation =
|
||||||
EncodeArgumentSelector(inputFile = fileResult.file, streams = streams, outFileName = outFileName)
|
EncodeArgumentSelector(collection = collection, inputFile = inFile, streams = streams, outFileName = outFileName)
|
||||||
produceEncodeMessage(baseMessage, encodeInformation.getVideoAndAudioArguments())
|
|
||||||
encodeInformation.getSubtitleArguments().forEach { s ->
|
val videoInstructions = encodeInformation.getVideoAndAudioArguments()
|
||||||
produceEncodeMessage(baseMessage, s)
|
if (videoInstructions == null) {
|
||||||
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_VIDEO, referenceId, "Failed to generate Video Arguments Bundle")
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
produceMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_VIDEO, Message(referenceId, Status(StatusType.SUCCESS)), videoInstructions)
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
fun createExtractWork(referenceId: String, collection: String?, inFile: String?, streams: MediaStreams?, outFileName: String?) {
|
||||||
|
if (inFile.isNullOrBlank()) {
|
||||||
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_SUBTITLE, referenceId, "No input file received"); return
|
||||||
|
}
|
||||||
|
if (streams == null) {
|
||||||
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_SUBTITLE, referenceId, "No input streams received"); return
|
||||||
|
}
|
||||||
|
if (outFileName.isNullOrBlank()) {
|
||||||
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_SUBTITLE, referenceId, "No output file name received!"); return
|
||||||
|
}
|
||||||
|
if (collection.isNullOrBlank()) {
|
||||||
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_SUBTITLE, referenceId, "No collection provided for file!"); return
|
||||||
|
}
|
||||||
|
|
||||||
|
val argsSelector = EncodeArgumentSelector(collection = collection, inputFile = inFile, streams = streams, outFileName = outFileName)
|
||||||
|
val items = argsSelector.getSubtitleArguments()
|
||||||
|
if (argsSelector == null || items.isEmpty()) {
|
||||||
|
produceErrorMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_SUBTITLE, referenceId, "Failed to generate Subtitle Arguments Bundle")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
argsSelector.getSubtitleArguments().forEach {
|
||||||
|
produceMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED_SUBTITLE, Message(referenceId, Status(StatusType.SUCCESS)), it)
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
private fun produceErrorMessage(baseMessage: Message, reason: String) {
|
final override fun loadDeserializers(): Map<String, IMessageDataDeserialization<*>> {
|
||||||
val message = Message(
|
return DeserializerRegistry.getEventToDeserializer(
|
||||||
referenceId = baseMessage.referenceId,
|
KafkaEvents.EVENT_READER_RECEIVED_FILE,
|
||||||
actionType = baseMessage.actionType,
|
KafkaEvents.EVENT_READER_RECEIVED_STREAMS,
|
||||||
Status(statusType = StatusType.ERROR, message = reason)
|
KafkaEvents.EVENT_READER_DETERMINED_FILENAME
|
||||||
)
|
)
|
||||||
messageProducer.sendMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED.event, message)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private fun produceEncodeMessage(baseMessage: Message, data: EncodeInformation?) {
|
|
||||||
val message = Message(
|
|
||||||
referenceId = baseMessage.referenceId,
|
|
||||||
actionType = baseMessage.actionType,
|
|
||||||
Status(statusType = if (data != null) StatusType.SUCCESS else StatusType.IGNORED),
|
|
||||||
data = data
|
|
||||||
)
|
|
||||||
messageProducer.sendMessage(KafkaEvents.EVENT_READER_ENCODE_GENERATED.event, message)
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
}
|
}
|
||||||
@ -1,8 +0,0 @@
|
|||||||
package no.iktdev.streamit.content.reader.analyzer.encoding.dto
|
|
||||||
|
|
||||||
data class EncodeInformation(
|
|
||||||
val inputFile: String,
|
|
||||||
val outFileName: String,
|
|
||||||
val language: String,
|
|
||||||
val arguments: List<String>
|
|
||||||
)
|
|
||||||
@ -11,15 +11,4 @@ class SubtitleEncodeArguments(val subtitle: SubtitleStream) {
|
|||||||
return result
|
return result
|
||||||
}
|
}
|
||||||
|
|
||||||
fun getFormatToCodec(): String? {
|
|
||||||
return when(subtitle.codec_name) {
|
|
||||||
"ass" -> "ass"
|
|
||||||
"subrip" -> "srt"
|
|
||||||
"webvtt", "vtt" -> "vtt"
|
|
||||||
"smi" -> "smi"
|
|
||||||
"hdmv_pgs_subtitle" -> null
|
|
||||||
else -> null
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
@ -1,16 +1,16 @@
|
|||||||
package no.iktdev.streamit.content.reader.analyzer.encoding.helpers
|
package no.iktdev.streamit.content.reader.analyzer.encoding.helpers
|
||||||
|
|
||||||
import no.iktdev.streamit.content.common.streams.AudioStream
|
import no.iktdev.exfl.using
|
||||||
import no.iktdev.streamit.content.common.streams.MediaStreams
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
import no.iktdev.streamit.content.common.streams.SubtitleStream
|
import no.iktdev.streamit.content.common.dto.reader.work.EncodeWork
|
||||||
import no.iktdev.streamit.content.common.streams.VideoStream
|
import no.iktdev.streamit.content.common.dto.reader.work.ExtractWork
|
||||||
|
import no.iktdev.streamit.content.common.streams.*
|
||||||
import no.iktdev.streamit.content.reader.analyzer.encoding.dto.AudioEncodeArguments
|
import no.iktdev.streamit.content.reader.analyzer.encoding.dto.AudioEncodeArguments
|
||||||
import no.iktdev.streamit.content.reader.analyzer.encoding.dto.EncodeInformation
|
|
||||||
import no.iktdev.streamit.content.reader.analyzer.encoding.dto.SubtitleEncodeArguments
|
import no.iktdev.streamit.content.reader.analyzer.encoding.dto.SubtitleEncodeArguments
|
||||||
import no.iktdev.streamit.content.reader.analyzer.encoding.dto.VideoEncodeArguments
|
import no.iktdev.streamit.content.reader.analyzer.encoding.dto.VideoEncodeArguments
|
||||||
import no.iktdev.streamit.content.reader.preference
|
import no.iktdev.streamit.content.reader.preference
|
||||||
|
|
||||||
class EncodeArgumentSelector(val inputFile: String, val streams: MediaStreams, val outFileName: String) {
|
class EncodeArgumentSelector(val collection: String, val inputFile: String, val streams: MediaStreams, val outFileName: String) {
|
||||||
var defaultSelectedVideo: VideoStream? = defaultSelectedVideo()
|
var defaultSelectedVideo: VideoStream? = defaultSelectedVideo()
|
||||||
var defaultSelectedAudio: AudioStream? = defaultSelectedAudio()
|
var defaultSelectedAudio: AudioStream? = defaultSelectedAudio()
|
||||||
|
|
||||||
@ -43,30 +43,40 @@ class EncodeArgumentSelector(val inputFile: String, val streams: MediaStreams, v
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
fun getVideoAndAudioArguments(): EncodeInformation? {
|
fun getVideoAndAudioArguments(): EncodeWork? {
|
||||||
val selectedVideo = defaultSelectedVideo
|
val selectedVideo = defaultSelectedVideo
|
||||||
val selectedAudio = getSelectedAudioBasedOnPreference() ?: defaultSelectedAudio
|
val selectedAudio = getSelectedAudioBasedOnPreference() ?: defaultSelectedAudio
|
||||||
return if (selectedVideo == null || selectedAudio == null) return null
|
return if (selectedVideo == null || selectedAudio == null) return null
|
||||||
else {
|
else {
|
||||||
EncodeInformation(
|
val outFileName = "$outFileName.mp4"
|
||||||
inputFile = inputFile,
|
val outFile = CommonConfig.outgoingContent.using(collection, outFileName)
|
||||||
outFileName = "$outFileName.mp4",
|
EncodeWork(
|
||||||
language = selectedAudio.tags.language ?: "eng",
|
collection = collection,
|
||||||
|
inFile = inputFile,
|
||||||
arguments = VideoEncodeArguments(selectedVideo).getVideoArguments() +
|
arguments = VideoEncodeArguments(selectedVideo).getVideoArguments() +
|
||||||
AudioEncodeArguments(selectedAudio).getAudioArguments()
|
AudioEncodeArguments(selectedAudio).getAudioArguments(),
|
||||||
|
outFile = outFile.absolutePath
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fun getSubtitleArguments(): List<EncodeInformation> {
|
fun getSubtitleArguments(): List<ExtractWork> {
|
||||||
return streams.streams.filterIsInstance<SubtitleStream>().map {
|
val subtitleStreams = SubtitleStreamSelector(streams.streams.filterIsInstance<SubtitleStream>())
|
||||||
val subArgs = SubtitleEncodeArguments(it)
|
|
||||||
EncodeInformation(
|
return subtitleStreams.getDesiredStreams().map {
|
||||||
inputFile = inputFile,
|
val args = SubtitleEncodeArguments(it)
|
||||||
outFileName = "$outFileName.${subArgs.getFormatToCodec()}",
|
val language = it.tags.language ?: "eng"
|
||||||
language = it.tags.language ?: "eng",
|
val outFileName = "$outFileName.${subtitleStreams.getFormatToCodec(it.codec_name)}"
|
||||||
arguments = subArgs.getSubtitleArguments()
|
val outFile = CommonConfig.outgoingContent.using(collection, "sub", language, outFileName)
|
||||||
|
|
||||||
|
ExtractWork(
|
||||||
|
collection = collection,
|
||||||
|
language = language,
|
||||||
|
inFile = inputFile,
|
||||||
|
outFile = outFile.absolutePath,
|
||||||
|
arguments = args.getSubtitleArguments(),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -0,0 +1,116 @@
|
|||||||
|
package no.iktdev.streamit.content.reader.collector
|
||||||
|
|
||||||
|
import kotlinx.coroutines.runBlocking
|
||||||
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
|
import no.iktdev.streamit.content.common.Downloader
|
||||||
|
import no.iktdev.streamit.content.common.SequentialKafkaReader
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.DeserializerRegistry
|
||||||
|
import no.iktdev.streamit.content.common.dto.Metadata
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.EpisodeInfo
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.FileResult
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.EncodeWork
|
||||||
|
import no.iktdev.streamit.library.db.query.*
|
||||||
|
import no.iktdev.streamit.library.db.tables.catalog
|
||||||
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.StatusType
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.sequential.SequentialMessageListener
|
||||||
|
import org.springframework.stereotype.Service
|
||||||
|
import java.io.File
|
||||||
|
|
||||||
|
@Service
|
||||||
|
class EncodedVideoConsumer: SequentialKafkaReader("collectorConsumerVideo") {
|
||||||
|
override val accept: KafkaEvents
|
||||||
|
get() = KafkaEvents.EVENT_READER_RECEIVED_FILE
|
||||||
|
override val subAccepts: List<KafkaEvents>
|
||||||
|
get() = listOf(
|
||||||
|
KafkaEvents.EVENT_METADATA_OBTAINED,
|
||||||
|
KafkaEvents.EVENT_READER_DETERMINED_SERIE,
|
||||||
|
KafkaEvents.EVENT_READER_DETERMINED_MOVIE,
|
||||||
|
KafkaEvents.EVENT_ENCODER_ENDED_VIDEO_FILE
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
val listener = object: SequentialMessageListener(
|
||||||
|
topic = CommonConfig.kafkaTopic,
|
||||||
|
consumer = defaultConsumer,
|
||||||
|
accept = accept.event,
|
||||||
|
subAccepts = subAccepts.map { it.event },
|
||||||
|
deserializers = loadDeserializers(),
|
||||||
|
validity = 86400000,
|
||||||
|
listener =this
|
||||||
|
) {}
|
||||||
|
|
||||||
|
init {
|
||||||
|
listener.listen()
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
override fun getRequiredMessages(): List<String> {
|
||||||
|
return listOf(accept.event) + subAccepts.map { it.event }
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun onAllMessagesProcessed(referenceId: String, result: Map<String, Message?>) {
|
||||||
|
val metadata = result[KafkaEvents.EVENT_METADATA_OBTAINED.event]?.data as Metadata?
|
||||||
|
val fileData = result[KafkaEvents.EVENT_READER_RECEIVED_FILE.event]?.data as FileResult?
|
||||||
|
val encodeStatus = result[KafkaEvents.EVENT_ENCODER_ENDED_VIDEO_FILE.event]?.status?.statusType ?: StatusType.ERROR
|
||||||
|
val encodeWork = result[KafkaEvents.EVENT_ENCODER_ENDED_VIDEO_FILE.event]?.data as EncodeWork?
|
||||||
|
|
||||||
|
|
||||||
|
if (fileData == null || encodeStatus != StatusType.SUCCESS || encodeWork == null) {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
val videoFileNameWithExtension = File(encodeWork.outFile).name
|
||||||
|
|
||||||
|
val contentType = metadata?.type ?: return
|
||||||
|
val iid = if (contentType == "movie") MovieQuery(videoFileNameWithExtension).insertAndGetId() else null
|
||||||
|
|
||||||
|
val serieData = result[KafkaEvents.EVENT_READER_DETERMINED_SERIE.event]?.data as EpisodeInfo?
|
||||||
|
if (serieData != null) {
|
||||||
|
val success = SerieQuery(serieData.title, serieData.episode, serieData.season, fileData.title, videoFileNameWithExtension).insertAndGetStatus()
|
||||||
|
if (!success)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
val coverFile = metadata?.cover?.let { coverUrl ->
|
||||||
|
runBlocking {
|
||||||
|
try {
|
||||||
|
Downloader(coverUrl, CommonConfig.outgoingContent, fileData.title).download()
|
||||||
|
} catch (e: Exception) {
|
||||||
|
// No cover
|
||||||
|
null
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
val metaGenre = metadata.genres
|
||||||
|
val gq = GenreQuery(*metaGenre.toTypedArray())
|
||||||
|
gq.insertAndGetIds()
|
||||||
|
val gids = gq.getIds().joinToString(",")
|
||||||
|
|
||||||
|
val cq = CatalogQuery(
|
||||||
|
title = fileData.title,
|
||||||
|
cover = coverFile?.name,
|
||||||
|
type = contentType,
|
||||||
|
collection = fileData.title,
|
||||||
|
iid = iid,
|
||||||
|
genres = gids
|
||||||
|
)
|
||||||
|
val cid = cq.insertAndGetId() ?: cq.getId() ?: return
|
||||||
|
if (!metadata.summary.isNullOrBlank()) {
|
||||||
|
val summary = metadata.summary ?: return
|
||||||
|
SummaryQuery(
|
||||||
|
cid = cid,
|
||||||
|
language = "eng", // TODO: Fix later,
|
||||||
|
description = summary
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
override fun loadDeserializers(): Map<String, IMessageDataDeserialization<*>> {
|
||||||
|
return DeserializerRegistry.getEventToDeserializer(*subAccepts.toTypedArray())
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,54 @@
|
|||||||
|
package no.iktdev.streamit.content.reader.collector
|
||||||
|
|
||||||
|
import kotlinx.coroutines.runBlocking
|
||||||
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
|
import no.iktdev.streamit.content.common.DefaultKafkaReader
|
||||||
|
import no.iktdev.streamit.content.common.Downloader
|
||||||
|
import no.iktdev.streamit.content.common.SequentialKafkaReader
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.DeserializerRegistry
|
||||||
|
import no.iktdev.streamit.content.common.dto.Metadata
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.FileResult
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.EncodeWork
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.work.ExtractWork
|
||||||
|
import no.iktdev.streamit.library.db.query.*
|
||||||
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
|
import no.iktdev.streamit.library.kafka.dto.StatusType
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.SimpleMessageListener
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.deserializer.IMessageDataDeserialization
|
||||||
|
import no.iktdev.streamit.library.kafka.listener.sequential.SequentialMessageListener
|
||||||
|
import org.apache.kafka.clients.consumer.ConsumerRecord
|
||||||
|
import java.io.File
|
||||||
|
|
||||||
|
class ExtractedSubtitleConsumer : DefaultKafkaReader("collectorConsumerSubtitle") {
|
||||||
|
|
||||||
|
private val listener = object: SimpleMessageListener(
|
||||||
|
topic = CommonConfig.kafkaTopic,
|
||||||
|
consumer = defaultConsumer,
|
||||||
|
accepts = listOf(KafkaEvents.EVENT_ENCODER_ENDED_SUBTITLE_FILE.event)
|
||||||
|
) {
|
||||||
|
override fun onMessageReceived(data: ConsumerRecord<String, Message>) {
|
||||||
|
val workResult = data.value().dataAs(ExtractWork::class.java)
|
||||||
|
if (!data.value().isSuccessful() || workResult == null) {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
val of = File(workResult.outFile)
|
||||||
|
SubtitleQuery(
|
||||||
|
title = of.nameWithoutExtension,
|
||||||
|
language = workResult.language,
|
||||||
|
collection = workResult.collection,
|
||||||
|
format = of.extension.uppercase()
|
||||||
|
).insertAndGetStatus()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
init {
|
||||||
|
listener.listen()
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun loadDeserializers(): Map<String, IMessageDataDeserialization<*>> {
|
||||||
|
return DeserializerRegistry.getEventToDeserializer(KafkaEvents.EVENT_ENCODER_ENDED_SUBTITLE_FILE)
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -9,6 +9,7 @@ import mu.KotlinLogging
|
|||||||
import no.iktdev.exfl.coroutines.Coroutines
|
import no.iktdev.exfl.coroutines.Coroutines
|
||||||
import no.iktdev.streamit.content.common.CommonConfig
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
import no.iktdev.streamit.content.common.Naming
|
import no.iktdev.streamit.content.common.Naming
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.FileResult
|
||||||
|
|
||||||
import no.iktdev.streamit.library.kafka.KafkaEvents
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
import no.iktdev.streamit.library.kafka.dto.Message
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
@ -106,10 +107,6 @@ class FileWatcher: FileWatcherEvents {
|
|||||||
messageProducer.sendMessage(KafkaEvents.EVENT_READER_RECEIVED_FILE.event , message)
|
messageProducer.sendMessage(KafkaEvents.EVENT_READER_RECEIVED_FILE.event , message)
|
||||||
}
|
}
|
||||||
|
|
||||||
data class FileResult(
|
|
||||||
val file: String,
|
|
||||||
val title: String = "",
|
|
||||||
val sanitizedName: String = ""
|
|
||||||
)
|
|
||||||
|
|
||||||
}
|
}
|
||||||
@ -5,6 +5,7 @@ import mu.KotlinLogging
|
|||||||
import no.iktdev.streamit.content.common.CommonConfig
|
import no.iktdev.streamit.content.common.CommonConfig
|
||||||
import no.iktdev.streamit.content.common.deamon.Daemon
|
import no.iktdev.streamit.content.common.deamon.Daemon
|
||||||
import no.iktdev.streamit.content.common.deamon.IDaemon
|
import no.iktdev.streamit.content.common.deamon.IDaemon
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.FileResult
|
||||||
import no.iktdev.streamit.content.reader.ReaderEnv
|
import no.iktdev.streamit.content.reader.ReaderEnv
|
||||||
import no.iktdev.streamit.content.reader.fileWatcher.FileWatcher
|
import no.iktdev.streamit.content.reader.fileWatcher.FileWatcher
|
||||||
import no.iktdev.streamit.library.kafka.KafkaEvents
|
import no.iktdev.streamit.library.kafka.KafkaEvents
|
||||||
@ -35,7 +36,7 @@ class StreamsReader {
|
|||||||
logger.info { "Ignoring event: ${data.key()} as status is not Success!" }
|
logger.info { "Ignoring event: ${data.key()} as status is not Success!" }
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
val dataValue = data.value().dataAs(FileWatcher.FileResult::class.java)
|
val dataValue = data.value().dataAs(FileResult::class.java)
|
||||||
|
|
||||||
if (dataValue == null) {
|
if (dataValue == null) {
|
||||||
logger.info { "Ignoring event: ${data.key()} as values is not of expected type!" }
|
logger.info { "Ignoring event: ${data.key()} as values is not of expected type!" }
|
||||||
@ -43,7 +44,7 @@ class StreamsReader {
|
|||||||
}
|
}
|
||||||
logger.info { "Preparing Probe for ${dataValue.file}" }
|
logger.info { "Preparing Probe for ${dataValue.file}" }
|
||||||
val output = mutableListOf<String>()
|
val output = mutableListOf<String>()
|
||||||
val d = Daemon(executable = ReaderEnv.ffprobe, parameters = listOf("-v", "quiet", "-print_format", "json", "-show_streams", dataValue.file), daemonInterface = object:
|
val d = Daemon(executable = ReaderEnv.ffprobe, daemonInterface = object:
|
||||||
IDaemon {
|
IDaemon {
|
||||||
override fun onOutputChanged(line: String) {
|
override fun onOutputChanged(line: String) {
|
||||||
output.add(line)
|
output.add(line)
|
||||||
@ -53,7 +54,7 @@ class StreamsReader {
|
|||||||
logger.info { "Probe started for ${dataValue.file}" }
|
logger.info { "Probe started for ${dataValue.file}" }
|
||||||
}
|
}
|
||||||
|
|
||||||
override fun onError() {
|
override fun onError(code: Int) {
|
||||||
logger.error { "An error occurred for ${dataValue.file}" }
|
logger.error { "An error occurred for ${dataValue.file}" }
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -63,7 +64,8 @@ class StreamsReader {
|
|||||||
|
|
||||||
})
|
})
|
||||||
val resultCode = runBlocking {
|
val resultCode = runBlocking {
|
||||||
d.run()
|
val args = listOf("-v", "quiet", "-print_format", "json", "-show_streams", dataValue.file)
|
||||||
|
d.run(args)
|
||||||
}
|
}
|
||||||
|
|
||||||
val message = Message(referenceId = data.value().referenceId, status = Status( statusType = if (resultCode == 0) StatusType.SUCCESS else StatusType.ERROR), data = output.joinToString("\n"))
|
val message = Message(referenceId = data.value().referenceId, status = Status( statusType = if (resultCode == 0) StatusType.SUCCESS else StatusType.ERROR), data = output.joinToString("\n"))
|
||||||
|
|||||||
@ -1,7 +1,7 @@
|
|||||||
package no.iktdev.streamit.content.reader.analyzer
|
package no.iktdev.streamit.content.reader.analyzer
|
||||||
|
|
||||||
|
import no.iktdev.streamit.content.common.deserializers.MediaStreamsDeserializer
|
||||||
import no.iktdev.streamit.content.common.streams.MediaStreams
|
import no.iktdev.streamit.content.common.streams.MediaStreams
|
||||||
import no.iktdev.streamit.content.reader.analyzer.encoding.EncodedDeserializers
|
|
||||||
import no.iktdev.streamit.library.kafka.consumers.DefaultConsumer
|
import no.iktdev.streamit.library.kafka.consumers.DefaultConsumer
|
||||||
import org.assertj.core.api.Assertions.assertThat
|
import org.assertj.core.api.Assertions.assertThat
|
||||||
import org.junit.jupiter.api.Assertions.*
|
import org.junit.jupiter.api.Assertions.*
|
||||||
@ -9,13 +9,12 @@ import org.junit.jupiter.api.Test
|
|||||||
|
|
||||||
class EncodedDeserializersTest {
|
class EncodedDeserializersTest {
|
||||||
|
|
||||||
val deserializer = EncodedDeserializers()
|
|
||||||
val consumer = DefaultConsumer.GsonDeserializer()
|
val consumer = DefaultConsumer.GsonDeserializer()
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun testDeserializationOfMediaStreams() {
|
fun testDeserializationOfMediaStreams() {
|
||||||
val message = consumer.deserialize("demo", messageMediaStream.toByteArray())
|
val message = consumer.deserialize("demo", messageMediaStream.toByteArray())
|
||||||
val result = deserializer.mediaStreams.deserialize(message)
|
val result = MediaStreamsDeserializer().deserialize(message)
|
||||||
assertInstanceOf(MediaStreams::class.java, result)
|
assertInstanceOf(MediaStreams::class.java, result)
|
||||||
assertThat(result?.streams).isNotNull()
|
assertThat(result?.streams).isNotNull()
|
||||||
assertThat(result?.streams).isNotEmpty()
|
assertThat(result?.streams).isNotEmpty()
|
||||||
|
|||||||
@ -1,6 +1,7 @@
|
|||||||
package no.iktdev.streamit.content.reader.streams
|
package no.iktdev.streamit.content.reader.streams
|
||||||
|
|
||||||
import com.google.gson.Gson
|
import com.google.gson.Gson
|
||||||
|
import no.iktdev.streamit.content.common.dto.reader.FileResult
|
||||||
import no.iktdev.streamit.content.reader.fileWatcher.FileWatcher
|
import no.iktdev.streamit.content.reader.fileWatcher.FileWatcher
|
||||||
import no.iktdev.streamit.library.kafka.dto.Message
|
import no.iktdev.streamit.library.kafka.dto.Message
|
||||||
import org.assertj.core.api.AssertionsForInterfaceTypes.assertThat
|
import org.assertj.core.api.AssertionsForInterfaceTypes.assertThat
|
||||||
@ -26,7 +27,7 @@ class StreamsReaderTest {
|
|||||||
""".trimIndent()
|
""".trimIndent()
|
||||||
assertDoesNotThrow {
|
assertDoesNotThrow {
|
||||||
val message = Gson().fromJson(data, Message::class.java)
|
val message = Gson().fromJson(data, Message::class.java)
|
||||||
val result = message.dataAs(FileWatcher.FileResult::class.java)
|
val result = message.dataAs(FileResult::class.java)
|
||||||
assertThat(result?.title).isEqualTo("Iseleve")
|
assertThat(result?.title).isEqualTo("Iseleve")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user