Profilbild von Anonymes Profil, Senior Expert Big Data Solution Architect & Enterprise Software Engineer
verfügbar

Letztes Update: 05.03.2024

Senior Expert Big Data Solution Architect & Enterprise Software Engineer

Abschluss: Master of Science (M.Sc.) in Business Informatics
Stunden-/Tagessatz: anzeigen
Sprachkenntnisse: deutsch (Muttersprache) | englisch (verhandlungssicher) | französisch (gut)

Dateianlagen

Maurice-Knopp-project-portfolio_040723.pdf

Skills

Current CV German: https://maknesium.de/projects/
Current CV English: https://maknesium.de/projects/index_en.html
  • Senior Expert Big Data Software Development (5+ Jahre)
  • Apache Spark | Apache Hive | Hadoop | Scala | Java | Python | Javascript | Hive Query Language (HQL) | SQL | NoSQL (MongoDB) | Apache Sqoop | Apache Airflow | Apache HBase | Apache Kafka | Gephi | ELK | Elasticsearch | Hazelcast | In Memory Solutions| Data Analysis | ETL

  • Senior Expert Enterprise Software Development (17+ Jahre)
  • Java 11 & 17 | OpenAPI REST | Spring Framework | Spring Boot | Quarkus Framework | Hibernate | jUnit | Oracle DB | PostgreSQL | MySQL | CoreMedia CMS | CoreMedia GSB | AngularJS | Solr | Kibana | Grafana | Microservices | Git | Jira | DevOps | SAP

  • Senior Expert Cloud Platforms (9+ Jahre)
  • AWS | Google GCP | Azure

  • DevOps (4+ Jahre)
  • Docker | Kubernetes | Helm | Gitlab Pipeline | Continuous Delivery | Continuous integration | CI/CD | GIT | Linux | Unix

Additional skills

Big Data
Microsoft Power BI
Tableau
CoreMedia
Teamfähigkeit
Data Science
Data Engineering
CMS
Microservices
Java
Scala
Java EE
JSP
Spring
Spring MVC
Spring Batch
Spring Boot
Java Servlets
Hibernate
Government Site Builder
Full-Stack Entwicklung
Eigeninitiative
Belastbarkeit
SQL
Python
Apache Hadoop
Apache Airflow
Apache Spark
Apache Hive
NoSQL
Apache Kafka
Hortonworks
Amazon AWS
Distributed Computing
AWS
Oracle DB
MySQL
PostgreSQL
Docker
Apache Solr
Apache Hbase
Gephi
elasticsearch
Gradle
OOP
OOD
Continuous Delivery
DevOps
Continuous Integration
IT-Projektmanagement
Ant
Maven
Bash
Javascript-Entwicklung
Linux Server
Linux
Ubuntu
Red Hat Enterprise Linux
CentOs
Webserver-Administration
Apache Tomcat
Nginx
Varnish
CDN
Eclipse
IntelliJ Idea
Subversion
Git
Videoverarbeitung
FFmpeg
C
C#
.NET Framework
Oracle Text
Apache Velocity
Apache Freemarker
RESTful Webservices
jQuery
Unix
Scrum
Cloud Computing
TypeScript
AngularJS
Self-Publishing
E-Publishing
E-Books (Amazon Apple Google)
Kindle Publishing
pandoc
LaTeX
Machine Learning
Blockchain
TensorFlow
Neuronale Netze
Deep Learning
Data Mining
iTunes Publishing
C/C++
JavaScript
Beratung
Teamleitung
SQL Server
Enterprise architect
Beratung
Visualisierung
Softwareentwicklung
Projektmanagement

Projekthistorie

08/2022 - bis jetzt
Solution Architect and data engineering for big data application
undisclosed company (Pharma und Medizintechnik, 1000-5000 Mitarbeiter)

Project: System and data architecture for big data application

Duration: ongoing (started in August 2022)
Role: Data architect, software architect for big data application solution
Size of team: 4 software developers, 10+ stakeholders representing functional requirements
Link(s): not public, internal software
Technologies used: Quarkus Framework, Camunda 8 Workflow Engine, OpenAPI, Junit 5, Hazelcast InMemory Grid, JDK17, Kubernetes, Helm, Docker, Confluence
Methodologies used: SCRUM
  • Worked on concept of data model for new software application handling billing of German governmental health insurance service
    • Translation of functional requirements into data model and suitable tech. Stack for software solution
    • Comparison of different technologies for processing and workflow control
  • Software design of in memory application and proof of concept implementation
  • Data architecture and persistence of historical data, microservice architecture
  • Implementation of proof of concepts with different technologies, measuring of access times to evaluate if solution would work on hundreds of millions of data points.

07/2013 - bis jetzt
Freiberuflich, Big Data Consulting & Enterprise Softwareentwicklung, Freiberuflich; ua. DuMont, Pixelpark (Endkunde: KfW), RTL, ]init[ (Endkunde: Bundesregierung), RheinEnergie
Freiberuflich

siehe: https://maknesium.de/projects/

10/2020 - 05/2023
Senior Java Enterprise Developer / REST-API Architect
undisclosed company (Energie, Wasser und Umwelt, 500-1000 Mitarbeiter)

Project: Relaunch of Web Customer Portal / OnlineService

Duration: ongoing (started in October 2020)
Role: Senior Java Enterprise Developer; REST-API software architect
Size of team: 4-6 Frontend & Backend Developers
Link(s): https://www.rheinenergie.com/de/ → OnlineService Link
Technologies used: Java 11, OpenAPI Spec, OpenAPI Generator, Liquibase, Quarkus, Oracle DB, SAP Open Data Protocol (Odata) API, wiremock, REST API, AngularJS, Gradle, Typescript, Docker, Kubernetes, KeyVault, Linux, GIT, GitLab, CI/CD, Kibana, AzureDevOps, Azure Arifacts, Amazon Cognito (IDP), REST-assured and Cypress tests, OWASP security check, microservice architecture, MuleSoft Anypoint
Methodologies used: SCRUM
 
  • Design, architecture and implementation of public high performance Java OnlineService API micro service as intermediary between the AngularJS frontend for customer facing data representation (OpenAPI) of the RE online services and the SAP ERP backend (Odata API) for customer authentication, storing customer and contract information
  • Backend serving up to 1 mio. customer requests for online service
  • Secured the micro service with state of the art security measures (OWASP) and implementation of API data validation to prevent fraud and data breach
  • Requirements engineering of business services with the help of skilled RE workers. Optimizing, implementing and verifying existing and new business services and functions within the java micro service.
  • Greatly improved service reliability by automatically monitoring metrics (AWS Kibana), implementing fixes and improvements for the software. Deployment of new software releases to production environment.
  • Developer discussions and peer review of source code to maintain high software quality
  • Implementation of frontend (cypress) and backend (postman REST lib) tests

05/2022 - 08/2022
Apache Spark and HBase Data Engineer and Java Software-Developer
undisclosed company (Pharma und Medizintechnik, 1000-5000 Mitarbeiter)

Project: Performance optimization of HBase apps, Spark job implementation

Duration: May 2022 – August 2022
Role: Data Engineer for performance optimization of HBase Access and Apache Spark
Size of team: 3 Java Developers
Link(s): not public, internal software
Technologies used: Apache Spark 3.1, Apache HBase 2.4.5 , Apache Kafka 3, Apache Hadoop 3.3; jUnit 5, fastutil big data performance library (https://fastutil.di.unimi.it)
Methodologies used: -
 
  • Optimizing performance for HBase access of complex data structures
    • identifying performance bottlenecks
    • optimizing access patterns and data processing flows
    • using fastutil big data performance library to reduce memory footprint and shorten run time of spark jobs processing of millions of data rows
  • Optimizing existing Spark jobs to reduce run time and memory footprint by 50%
  • Implementing Spark jobs accessing HBase data structures and Kafka messaging systems to meet functional requirements

08/2018 - 03/2020
Data Engineer + Principal Java Enterprise Software Developer
undisclosed company (Medien und Verlage, 1000-5000 Mitarbeiter)

Project: Graph generation based on massive data-set (~40 billion daily data points)

Duration: 20 months (August 2018 – March 2020)
Role: Data Engineer + Java Enterprise Software Developer
Size of team: 2 Big Data Java Software-Developers and 1 Data Engineer, 1 Data Scientist
Link(s): https://www.ip.de/online/news/cross_device_kurz_erklaert.cfm
Technologies used: Java 11, Scala 2.11, Hortonworks HDP 2.6/3.1: Apache Hadoop, Apache Spark, Apache Hive, Apache Hbase, Apache Kafka, HDFS, TEZ, Zookeeper, AWS S3, SQL/HQL, Python, Linux, DevOps, ClickHouse DB, PostgreSQL, Bash, Gephi, Apache Airflow, Apache Sqoop
Methodologies used: SCRUM
 
  • Exploration of source data: Planning and designing a data pipeline solution to clean and process massive data-set (~40 billion data points per day) to form a graph structure for easy clustering of nodes
  • Building a scalable java application (Framework: Spring Boot) to build up a graph structure with (weighted) edges between vertices/nodes.
    • Main challenge is to build up graph structure in memory from massive data-set (50+ Gigabytes per day)
    • Graph traversal has super-fast to identify known nodes when a new event is analysed → only in memory solution is feasible
    • Implementation of solution on premise at customer’s data center (running Hortonworks distribution 2.6 and later newer 3.1)
    • Development of algorithms handling ~700 GB of in memory data,
    • Handling fail-proof connection to HDFS data lake, ORC files on HDFS contain source data
    • Exporting (sub-)graph structures in different formats for ad targeting engine(s) of international media companies
    • Continuously measuring and evaluating graph quality based on heuristics since no labeled data exists for given real-world data
    • Technology was used to run several large ad campaigns
  • Visualization of data for insight generation for marketing department
  • Git workflow; merge-requests signed off by team members
  • setup and use of CI/CD pipeline
  • TDD – test driven development based on JUnit for core algorithms
  • Size of Hadoop cluster: 20 nodes with 300 TB HDFS storage and 1.2 TB of memory running Hortonworks HDP 2.6

Reisebereitschaft

Verfügbar in den Ländern Deutschland, Österreich und Schweiz
D-A-CH + Remote

Sonstige Angaben

Current CV German: https://maknesium.de/projects/ Current CV English: https://maknesium.de/projects/index_en.html
Profilbild von Anonymes Profil, Senior Expert Big Data Solution Architect & Enterprise Software Engineer Senior Expert Big Data Solution Architect & Enterprise Software Engineer
Registrieren