Projects using Matplotlib
Projects using Matplotlib
Sign Up
Post a job
Sign Up
Log In
Filters
1
Projects
People
0
Venkata Anirudh Parakala
Geographical Understanding of Twitter Health Topics using NLP
0
7
0
Uthara
Allergy Prediction Model Development
0
11
0
Aryan Kushwah
NASDAQ Momentum Trading Strategy
0
7
0
Aaron Segiel
CIFAR-10 Image Classification with CNN
0
8
0
Anastasiya Kotelnikova
Spiking Neural Networks with PyTorch & Norse
0
4
0
Júlio Silva
Movies Data Scraping and Analysis
0
10
3
Ifigeneia Tsiflidou
Revenue Prediction Model Development
3
25
0
Sharles Maurício Mariano
Personality Type Mapping with Machine Learning
0
5
0
Ravi Mahawar
Fractal Image Generation with OpenMP and MPI
0
5
0
Eyad Gad
Image Processing and Analysis of the Scratch Assay
0
3
0
Persia Cooper
Netflix Data Analysis
0
1
0
hussein hafez
Fraud Detection on Credit Card Transactions
0
4
0
Uchenna Ejike
Global Power Plant Database Analysis
0
3
0
Ismael Charpentier
Hopefield Neural Network
0
10
2
Yusuf Adetona
1M+ Row Migration ROI: SQL Server to PostgreSQL I engineered a Python ETL pipeline to migrate legacy SQL Server data to PostgreSQL, prioritizing cost-efficiency and absolute integrity. The Problem: High licensing costs and a need for cloud-scalability without risking the "Source of Truth." The Outcomes: 🔹Savings: Projected 65% reduction in annual licensing fees 🔹Integrity: 1,055,008 rows reconciled with zero violations 🔹Audit: Flagged 24,000+ quality issues (negative prices, invalid emails) during the move The Impact: Shifted from reactive maintenance to proactive cloud-scaling on AWS/Azure with a cleaner, more reliable database. Why Me? As an Accountant & Data Engineer, I bring "ledger-style" reconciliation to architecture. I don’t just move data; I audit it. Open for Migration contracts. Let’s cut your costs.
2
40
1
Molka B
AI Bias Detection & Mitigation Tool for Hiring Models My role: AI/ML Engineer & Fairness Auditor Project description: AI Bias Detection and Mitigation Framework for Recruitment Systems. Built a complete, production-ready tool that: Detects gender, race, and age bias in hiring ML models Uses industry-standard fairness metrics (Disparate Impact, Demographic Parity, Equal Opportunity) Applies Reweighing mitigation (AIF360) → improves fairness while keeping accuracy loss < 2% Includes interactive Gradio demo for live testing Full business impact section + compliance notes (EEOC 4/5th rule, GDPR, EU AI Act)
1
68
Explore projects