diff --git a/index.html b/index.html
index 0949a23..dbbbc26 100644
--- a/index.html
+++ b/index.html
@@ -258,7 +258,7 @@
Technical competencies
Database Management: |
- PostgreSQL, SQLite3, Neo4j with expertise in SQLAlchemy
+ PostgreSQL, SQLite3, Neo4j, SQLAlchemy
|
@@ -272,14 +272,14 @@ Technical competencies
Data Science: |
- Pandas, PyMC3, Scikit-Learn, sktime, Seaborn
+ Pandas, PyMC3, scikit-Learn, sktime, Seaborn
|
Data Engineering: |
- Kedro, prefect, PySpark
+ Kedro, Prefect, PySpark
|
@@ -333,7 +333,7 @@ Freelance projects (Oct 2022-present)
- Automated SQL Script Generation for Cross-Platform Data Migration
+ Automated SQL Script Generation for Cross-Platform Data Migration in PostgreSQL
@@ -428,11 +428,11 @@
- - Improved performance of information retrieval by 20% on unseen test data using a custom named entity recognition (NER) from Spacy.
+ - Improved performance of information retrieval by 20% on unseen test data using a custom named entity recognition (NER) from Spacy.
- - Performed POC’s on Azure DataBricks environment to improve model performance using rule-based techniques as well as NER and annotated data to train custom NER.
+ - Performed POC’s on Azure DataBricks environment to improve model performance using rule-based techniques as well as NER and annotated data to train custom NER.
- - Added text preprocessing features to the NLP pipeline such as spacy tokenization, Part of speech (POS) tagging, better handling of non‑english emails, breaking emails into sentences, etc.
+ - Added text preprocessing features to the NLP pipeline such as Spacy tokenization, Part of speech (POS) tagging, better handling of non‑english emails, breaking emails into sentences, etc.
@@ -482,7 +482,7 @@
- Customer Promotional Responsiveness modeling by marketing channel
+ Python Framework for Customized Promotional Responsiveness Models Across Regions
@@ -503,15 +503,15 @@
- - Developed a Python package that abstracts the complexities of the data science workflow, enabling configurable deployments across diverse scenarios such as different countries and disease areas
+ - Developed a Python package with Cookiecutter templates that abstract the complexities of the data science workflow, enabling configurable deployments across diverse scenarios such as different countries and disease areas.
- Enhanced the package to seamlessly wrap over scikit-learn, thereby simplifying key data science tasks from preprocessing to model training and tuning
- - Incorporated MLflow into the package for robust artifact management, allowing for the tracking of model versions, data inputs, and predictions
+ - Incorporated MLflow into the package for robust artifact management, allowing for the tracking of model versions, data inputs, and predictions
- Created customer segmentation models and proposed optimal resource allocation based on customer responsiveness to different marketing channels
- - Investigated adaptations to data science methodology for country/product specificities for maximum reusability. Delivered as many as ten different use cases as lead data scientist for different products and countries
+ - Investigated adaptations to data science methodology for country/product specificities for maximum reusability. Delivered as many as ten different use cases for different products and countries
- Supported data engineers in the creation of features using PySpark and validated ingested data using data visualization methods and discussions with subject-matter experts
@@ -637,11 +637,11 @@
- Developed web scraper using Beautiful Soup to collect information such as apartment data such as price, area, etc.
- - Implemented SQLite for data storage, using `pydantic` for data validation and `SQLalchemy` for database interactions.
+ - Implemented SQLite for data storage, using `Pydantic` for data validation and `SQLAlchemy` for database interactions.
- Encapsulated the concerns into a python package with dependency management using Poetry.
- - Utilized Prefect for task scheduling, ensuring monitoring of data collection.
+ - Employed Prefect for job orchestration, managing the workflow's scheduling and monitoring of scraping tasks.