Tegisty Hailay profile site
Tegisty Hailay Degef
Harar, Ethiopia
Data Engineering
Apache Framework(Spark, Kafka and Airflow
ETL/ELT and DBT
Tableau and Power BI
PostgreSQL
SQL
programing Lnguage
Python
Java and java script
C++
PHP and Laravel
Machine Learming Tools
Tensorflow, Scikit-learn, Keras
Matplotlib, Seaborn
Pandas, Numpy
PyTorch
CI/CD & Devops
CI/CD
Github Actions
CML
Kubernetes
Docker
Operating System &technologies
Unix/Linux
Window
AWS
Math and stat
Linear Algebra
Applied Mathematics I, II &III
Advanced Calculus
About me
A junior Data Engineer with computer science background who can build a scalable data pipeline using cloud and data engineering technologies for supporting machine learning experiments, data science models, business intelligence reporting and data exchange with partners. Additional skill in ML, DL, NLP, visualization, modelling, MLOPs and CI/CD.
Education
Bule-Hora University ( 2015-2019 )
BSc.(Computer Science)
Software Engineering
C/C++ Programming
Data structure and algorithm analysis
Artificial intelligence / Machine learning
Java and Object-oriented programing
Database (SQL, Oracle, MYSQL)
Descriptions of key projects: Automatic Timetable Generation System
10Academy Training ( 15 Aug 2022- 15 Nov 2022 )
Data Engineer, Machine Learning and Web3 Engineer
Designing Data Pipeline for entire ELT project
DBT transformation
Data Loading from Azure to snowflake and prepare for visualization
DBT test on models
Working and deploying in Azure Devops/AWS
Application and System Developer (Internship)
Mekelle University (May 2018- September 2018)
Role: Team Leader
projects: Automatic timetable generation System and Notice Board
Content, as per Bock's law
Projects
Text-to-Speech data collection with Kafka, Airflow, and Spark
A project to produce a tool that can be deployed to process posting and receiving text and audio files from and into a data lake, apply transformation in a distributed manner, and load it into a warehouse in a suitable format to train a speech-to-text model.
Scalable Data pipeline warehouse with airflow, DBT and Postgres
A project to create a scalable data warehouse that host the vehicle trajectory data extracted by analyzing footage taken by swarm drones and static roadside cameras. ELT framework along with airflow used to setup transformation workflows on the required objective. Airflow for scheduling tasks and DBT for transformation and build data model for analytics purpose was applied
Rosman Pharmaticula Sales Prediction
The main task of this project was to predict daily sales in various stores up to 6 weeks ahead of time. Deep Learning techniques was used to predict various outcomes including but not limited to future sales. From the analysis I found that promotion has an impact on sales.
Computer Vision for Creative Optimization: KPI maximization through image analysis
The main task of this project is to apply deep learning based computer vision techniques for creative optimization in mobile advertising. Different deep learning based image analysis (feature extraction and segmentation) techniques were used and a random forest regression algorithm was applied to predict the KPI performance.