Welcome to my Data Science and Generative AI portfolio. This workspace is organized to showcase end-to-end projects, from SQL data extraction and Exploratory Data Analysis (EDA) to Deep Learning and Generative AI implementations.
Projects/
├── data/ # Raw and processed datasets
├── notebooks/ # Jupyter notebooks sorted by project type
├── sql/ # SQL scripts, schemas, and queries
├── src/ # Production-grade source code
├── projects/ # Dedicated folders for individual projects
├── models/ # Saved model weights and binaries
├── reports/ # Analysis and project reports
├── dashboards/ # Visualization dashboard files (PowerBI/Tableau)
├── configs/ # Configuration files
└── tests/ # Unit and integration tests
I have provided a setup_env.bat script to automate the environment creation.
-
Run the setup script:
setup_env.bat
This will create a
venv, upgrade pip, and install all dependencies. -
Manual Activation:
venv\Scripts\activate
If you prefer manual installation:
pip install -r requirements.txtTo explore the analysis and models:
jupyter notebookNavigate to the notebooks/ folder.
Pre-trained models are stored in models/. To run inference scripts:
python projects/03-ml-project/predict.pyInteractive dashboards:
streamlit run src/visualization/app.pyTo serve models via API:
uvicorn src.main:app --reloadSQL scripts and schema definitions are located in the sql/ directory. You can execute them using your preferred SQL client or via the sqlalchemy scripts in src/data/.
- Core DS: Pandas, NumPy, Scikit-Learn, Matplotlib, Seaborn
- Deep Learning: PyTorch, TensorFlow
- Generative AI: Transformers (HuggingFace), OpenAI, LangChain
- SQL: SQLite, MySQL, SQLAlchemy
- Web App: Streamlit, FastAPI
Ravi Soni