Générale Technical development leader of a regulatory project -Mesh contra ct-to address the IFRS 9.2 requirements in term of regulations using Big Data technologies at Société Générale:
▪Hortonworks consultant
▪Defining the software stack for the project.
▪Contributing and leading the developments : Mesh Contrat relies on many technical components: Oozie, spark (scala), spark streaming, kafka, Teradata, sqoop, elasticsearch and kibana.
▪Implementing the continuous delivery/integration process for the project ith Nexus, Jenkins, ansible
▪Successful production deployment of the project Hortonworks Solution Architect at Société Générale:
Hadoop (Hortonworks):
▪Hadoop Security Expert: Designing and implementing of secured solutions for security requirements.
▪Installation and configuration of a new secured development/integration cluster for projects with ranger and Kerberos enabled.
▪Synchronization ranger, with LDAPs, and Configuring sssd for ldap authentication
▪Full automation of installation and configuration of components/products for the cluster with ansible
▪Configuring backup cluster, and providing solutions for disaster recovery strategies.
▪Configuring and running mirror-maker to backup streaming data in secured environments (Kafka Acls; SSL and Kerberos).
▪Defining and implementing the migration strategy from using Kafka ACLs
to Ranger policies and migration from self -signed certificates to CA signed certificates for Kafka SSL listener.
▪Enabling wire encryption and managing SSL certificates on major Hadoop components.
▪Installing and configuring Hue on a HA and kerberized cluster and synchronization with ldap.
▪Installing and configuring Knox to connect reporting tools on Hive such as Tableau.
▪Setup of Prometheus for monitoring and alerting of the most critical components: ldap, FS size, ...
Talend:
▪Define and implementation in all Societe Genreale environments.
▪Connecting the different TAC instances to the Active Directory group and Securing the communication with SSL.
▪Implementing ansib le playbooks to install TAC and jobservers.
▪Define and implementation the logging strategy for Talend projects that use Kafka (SASL)
▪Defining best practices and security strategies to isolate jobservers
with cgroups for projects and to authenticate each jobserver with Kerberos.
▪Configuration and installation of Talend Data Quality on a kerberized environment: Integration with Kafka for data dictionary service and HDFS to import/export data.
▪HDPA trainer at Ysance: Administration + Security + Preparing for HDP administrator certification.
▪Trainer at Canal+ on how to setup Cloudbreak on AWS.
▪Trainer at canal+ on the full HDP stack
▪Hortonworks Administrator certified (HDPCA)
Data Architect /Developer at SFR July 2016 -October 2016 January 2016 July 2016
Poc design and implementation of a monitoring solution based on big data technologies for SFR:
▪Poc design and implementation of a monitoring solution for the VoD plateform of SFR.
▪Specification of the KPIs to be monitored with the customer.
▪Building from scratch and securing the monitoring plateform on AWS. The plateform relies on the following components: logstash, S3, elasticsearch,
EMR, spark, kibana, nginx, SSL.
▪Implementing the logstash configuration to parse and normalize the logs, and the spark jobs to generate the business views and index them in Elasticsearch.
▪The poc was enough mature to run for around 1 year on AWS without any
issue and has triggered a new project at SFR to internalize the poc.
Data Architect/Developer at BNP Solution architect:
▪Definition and implementation of a new monitoring and alerting solution
based on HDP and influxDataPlateform (Telegraf, kapacitor, influxDb)stacks to monitor more than 500 VM+ databases.
▪Automation of the deployment of the monitoring stack with puppet and saltstack.
▪Configure all Telegraf agents to send telemetry data to kafka. Then Implementation of a spark streaming job to consume data from kafka, commit the offsets to Zookeeper for roll back and load data into InfluxDb.
▪Automation of deployment of HDP clusters
using blueprints, saltstak on VMs.
▪Poc: configuration and installation of HDFS transparency connector to plug
GPFS-IBM on yarn and execution of spark jobs on GPFS.
▪Definition of the indexing strategies on Elasticsearch on a 84 node cluster.
▪PHD on recommendation systems and video caching algorithms.
▪Orange Data traffic analysis to derive user behavior patterns to fine tune the recommendation system.
▪4 publications on top computer of science conferences
▪Participation to 2 European projects: ecousin/ocean