Skip to content

Commit bec29be

Browse files
Add preliminary output tables and security policy documentation
- Created CSV files for various models' performance metrics: - Actor-Critic Reward Table - Bayesian Decision-Making Accuracy Table - Logistic Regression Accuracy Table - Multi-Arm Bandit Reward Table - Neural Network Training Loss Table - Policy Gradient Reward Table - Q-Learning Reward Table - SVM Accuracy Table - Added SECURITY.md to outline the security policy, reporting vulnerabilities, and dependency security practices. - Introduced requirements.txt to specify project dependencies.
1 parent bd73649 commit bec29be

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

44 files changed

+4097
-183
lines changed
Lines changed: 38 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,38 @@
1+
---
2+
name: Bug report
3+
about: Create a report to help us improve
4+
title: ''
5+
labels: ''
6+
assignees: ''
7+
8+
---
9+
10+
**Describe the bug**
11+
A clear and concise description of what the bug is.
12+
13+
**To Reproduce**
14+
Steps to reproduce the behavior:
15+
1. Go to '...'
16+
2. Click on '....'
17+
3. Scroll down to '....'
18+
4. See error
19+
20+
**Expected behavior**
21+
A clear and concise description of what you expected to happen.
22+
23+
**Screenshots**
24+
If applicable, add screenshots to help explain your problem.
25+
26+
**Desktop (please complete the following information):**
27+
- OS: [e.g. iOS]
28+
- Browser [e.g. chrome, safari]
29+
- Version [e.g. 22]
30+
31+
**Smartphone (please complete the following information):**
32+
- Device: [e.g. iPhone6]
33+
- OS: [e.g. iOS8.1]
34+
- Browser [e.g. stock browser, safari]
35+
- Version [e.g. 22]
36+
37+
**Additional context**
38+
Add any other context about the problem here.

.github/dependabot.yml

Lines changed: 11 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,11 @@
1+
version: 2
2+
updates:
3+
- package-ecosystem: "pip"
4+
directory: "/"
5+
schedule:
6+
interval: "weekly"
7+
open-pull-requests-limit: 5
8+
assignees:
9+
- TechMLW
10+
commit-message:
11+
prefix: "deps"

LICENSE

Lines changed: 21 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,21 @@
1+
MIT License
2+
3+
Copyright (c) 2026 Palash Dave
4+
5+
Permission is hereby granted, free of charge, to any person obtaining a copy
6+
of this software and associated documentation files (the "Software"), to deal
7+
in the Software without restriction, including without limitation the rights
8+
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9+
copies of the Software, and to permit persons to whom the Software is
10+
furnished to do so, subject to the following conditions:
11+
12+
The above copyright notice and this permission notice shall be included in all
13+
copies or substantial portions of the Software.
14+
15+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16+
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17+
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18+
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19+
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20+
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21+
SOFTWARE.

README.md

Lines changed: 120 additions & 123 deletions
Original file line numberDiff line numberDiff line change
@@ -1,133 +1,130 @@
1-
# Quantum Machine Learning for Strategic Decision Making
1+
# Machine Learning & Reinforcement Learning Algorithms
22

3-
This repository is currently under developement for implementation of quantum machine learning on strategic decision making for optimal output
3+
![Python](https://img.shields.io/badge/Python-3.8%2B-blue?logo=python)
4+
![Jupyter](https://img.shields.io/badge/Jupyter-Notebook-orange?logo=jupyter)
5+
![License](https://img.shields.io/badge/License-MIT-green)
46

5-
## Stage 1 — Learn Only the Needed Quantum Basics
7+
This repository contains a collection of Jupyter notebooks implementing fundamental machine learning, decision-making, and reinforcement learning algorithms.
8+
Each notebook focuses on a specific concept, combining theory with practical implementation for educational and experimental purposes.
69

7-
- **Goal**: Understand enough quantum computing to implement QML.
10+
---
811

9-
*Day 1–2*: Minimum QC Theory
12+
## 📂 Repository Structure
1013

11-
Everyone learns:
14+
code
15+
├── Research Models
16+
│ ├── multi-arm-bandit.ipynb
17+
│ ├── q-learning.ipynb
18+
│ ├── svm.ipynb
19+
├── actor-critic.ipynb
20+
├── bayesian-decision-making.ipynb
21+
├── logistic-regression.ipynb
22+
├── multi-arm-bandit.ipynb
23+
├── neural-network.ipynb
24+
├── policy-gradient.ipynb
25+
├── q-learning.ipynb
26+
└── svm.ipynb
1227

13-
- Qubits & superposition
14-
- Basic gates (X, Y, Z, H, CNOT, Rotation gates)
15-
- Measurement
16-
- Quantum circuits
28+
---
1729

18-
Use:
19-
- Qiskit textbook (free)
20-
- PennyLane “Intro to Quantum” tutorials
30+
## Research Paper
2131

22-
*Day 3–4*: What is QML?
32+
The following algorithms are specifically used in the research paper:
2333

24-
Learn:
25-
- Angle embedding
26-
- Amplitude embedding
27-
- Variational Quantum Circuits (VQC)
28-
- Quantum Neural Networks (QNNs)
29-
- Hybrid quantum-classical training
34+
1. [Support Vector Machine (SVM)](https://github.com/TechMLW/QuantFP/blob/master/code/svm.ipynb)
35+
2. [Multi-Armed Bandit](https://github.com/TechMLW/QuantFP/blob/master/code/multi-arm-bandit.ipynb)
36+
3. [Q-Learning](https://github.com/TechMLW/QuantFP/blob/master/code/q-learning.ipynb)
3037

31-
Watch:
32-
- Xanadu PennyLane QML crash course
33-
- Qiskit ML playlist
34-
35-
*Day 5–7*: Divide Topics for Literature Review
36-
Each member summarizes 3–4 papers.
37-
38-
Topics:
39-
- QML basics
40-
- VQC for classification
41-
- QRL (Quantum Reinforcement Learning)
42-
- Quantum advantage claims
43-
- Applications in strategy: finance, optimization, multi-agent systems
44-
45-
Deliverables by end of Week 1:
46-
✔ Literature survey
47-
✔ Understand VQC models
48-
✔ Choose QML framework: PennyLane (recommended)
49-
50-
*Week 2* — Build QML Foundations
51-
52-
- **Goal**: Implement small QML models so your team becomes comfortable.
53-
54-
Choose one framework:
55-
- PennyLane + PyTorch (easiest)
56-
- Qiskit Machine Learning
57-
58-
*Day 8–10*: Everyone Implements a Simple QML Classifier
59-
60-
Example tutorial to follow:
61-
- PennyLane “Variational Classifier”
62-
- Build model on Iris dataset
63-
- Team Member B leads.
64-
65-
*Day 11–13*: Play With More Complex QML Models
66-
67-
Optional models:
68-
- Quantum Kernel-based SVM
69-
- QAOA for optimization
70-
- VQC for regression
71-
72-
*Day 14*: Decide the Final Experiment Setup
73-
74-
Pick ONE for your research experiment:
75-
- Option A: QML for Game-Theoretic Decisions
76-
- Predict opponent action
77-
- Quantum classifier for rock-paper-scissors strategy
78-
- Option B: QML for Multi-Armed Bandit Decisions (Best choice)
79-
- Use a VQC to predict best arm
80-
- Compare with classical models
81-
- wShow regret curves
82-
- Option C: QML for Supply Chain / Pricing decision
83-
- Classify optimal pricing or inventory action
84-
- Team D finalizes choice and dataset.
85-
86-
Deliverables by end of Week 2:
87-
✔ Working QML models
88-
✔ Classical ML baseline models ready
89-
✔ Final idea + dataset/environment chosen
90-
91-
*Week 3* — Implement the Strategic Decision-Making Model
92-
93-
- **Goal**: Apply QML to your decision-making task.
94-
95-
*Day 15–17*: Build Problem Environment
96-
97-
Examples:
98-
- If multi-armed bandit:
99-
- Simulate reward distributions
100-
- Encode context as angles
101-
- Use QML model to choose arm
102-
103-
If game theory:
104-
- Encode opponent behavior vectors
105-
- Train QML model to classify best response
106-
107-
*Day 18–20*: Train & Evaluate QML Model
108-
109-
Train with gradient descent
110-
111-
Use 4–8 qubits only (for simulators)
112-
Track:
113-
- accuracy
114-
- reward
115-
- regret
116-
- stability
117-
118-
*Day 21*: Compare with Classical Models
119-
120-
Team C leads baseline models:
121-
- Logistic regression
122-
- Shallow neural network
123-
- SVM
124-
125-
Make comparison plots:
126-
- Accuracy vs iterations
127-
- Reward vs episodes
128-
- Decision quality
129-
130-
Deliverables by end of Week 3:
131-
✔ Fully trained QML and classical models
132-
✔ Performance comparison
133-
✔ Preliminary results
38+
---
39+
40+
## 📘 Notebook Descriptions
41+
42+
1. [Logistic Regression](https://github.com/TechMLW/QuantFP/blob/master/code/logistic-regression.ipynb)
43+
- Binary classification using logistic regression
44+
- Model formulation, training, and evaluation
45+
-Gradient-based optimization
46+
47+
2. [Support Vector Machine (SVM)](https://github.com/TechMLW/QuantFP/blob/master/code/svm.ipynb)
48+
- Linear and margin-based classification
49+
- Decision boundaries and hinge loss
50+
- Practical implementation from scratch
51+
52+
3. [Neural Network](https://github.com/TechMLW/QuantFP/blob/master/code/neural-network.ipynb)
53+
- Feedforward neural network implementation
54+
- Activation functions and backpropagation
55+
- Training and inference workflow
56+
57+
4. [Bayesian Decision Making](https://github.com/TechMLW/QuantFP/blob/master/code/bayesian-decision-making.ipynb)
58+
- Probabilistic reasoning under uncertainty
59+
- Bayesian inference and decision rules
60+
- Applications to optimal decision policies
61+
62+
5. [Multi-Armed Bandit](https://github.com/TechMLW/QuantFP/blob/master/code/multi-arm-bandit.ipynb)
63+
- Exploration vs. exploitation trade-off
64+
- ε-greedy and related strategies
65+
- Performance comparison of bandit algorithms
66+
67+
6. [Q-Learning](https://github.com/TechMLW/QuantFP/blob/master/code/q-learning.ipynb)
68+
- Model-free reinforcement learning
69+
- Q-table updates and temporal-difference learning
70+
- Policy derivation from learned values
71+
72+
7. [Policy Gradient](https://github.com/TechMLW/QuantFP/blob/master/code/policy-gradient.ipynb)
73+
- Direct policy optimization methods
74+
- Stochastic policies and gradient estimation
75+
- Reinforcement learning with function approximation
76+
77+
8. [Actor-Critic](https://github.com/TechMLW/QuantFP/blob/master/code/actor-critic.ipynb)
78+
- Hybrid value-based and policy-based approach
79+
- Actor and critic architecture
80+
- Advantage estimation and learning stability
81+
82+
---
83+
84+
## 🛠️ Requirements
85+
86+
To run the notebooks, ensure the following are installed:
87+
88+
- Python 3.8+
89+
- Jupyter Notebook / JupyterLab
90+
- NumPy
91+
- Pandas
92+
- Matplotlib
93+
- (Optional) SciPy, scikit-learn
94+
95+
Install dependencies using:
96+
97+
```bash
98+
pip install numpy matplotlib scikit-learn jupyter
99+
```
100+
101+
---
102+
103+
## ▶️ How to Run
104+
105+
1. Clone the repository:
106+
107+
```bash
108+
git clone https://github.com/TechMLW/QuantFP
109+
cd QuantFP
110+
```
111+
112+
2. Launch Jupyter Notebook:
113+
114+
```bash
115+
jupyter notebook
116+
```
117+
118+
3. Open any .ipynb file and run the cells sequentially.
119+
120+
---
121+
122+
## 🎯 Purpose
123+
124+
This repository is intended for:
125+
126+
- Learning core machine learning and reinforcement learning algorithms
127+
- Academic coursework and self-study
128+
- Experimentation with algorithmic concepts from scratch
129+
130+
---

0 commit comments

Comments
 (0)