Each stage is evaluated separately (0–100 points), focusing on:

STAGE 1 – PROJECT PLANNING & DESIGN

Goal: Assess students’ ability to define a problem, design a neural network approach, and plan the implementation effectively.

Criteria Description Performance Levels Max Points
1. Problem Definition & Objectives The ML problem is clearly identified with defined input/output and measurable goals. ☐ Unclear problem (0–25)
☐ Defined but lacks measurable metrics (26–50)
☐ Clear objectives with evaluation metrics (51–75)
☐ Well-structured, quantifiable, and practical objectives (76–100)
25
2. Data Collection & Description Identifies data sources, data size, key features, and data quality issues. ☐ No data plan (0–25)
☐ Basic description only (26–50)
☐ Includes statistics or examples (51–75)
☐ Detailed EDA with visuals and justification (76–100)
25
3. Model & Pipeline Design Proposes suitable NN architecture (MLP, CNN, RNN…) and defines the workflow from preprocessing to evaluation. ☐ Missing or incorrect (0–25)
☐ Basic structure (26–50)
☐ Logical design with diagrams (51–75)
☐ Well-justified architecture and performance expectations (76–100)
25
4. Implementation Plan & Tools Defines timeline, task assignments, frameworks (TensorFlow, PyTorch, etc.), and computational resources. ☐ Not defined (0–25)
☐ Partial timeline (26–50)
☐ Detailed plan with clear tool selection (51–75)
☐ Comprehensive plan with risk management (76–100)
25
    Stage 1 Total /100

1. Problem Definition & Objectives (25 points)

What to include:

Examples:

Tips:

2. Data Collection & Description (25 points)

Include:

Examples:

Tips:

3. Model & Pipeline Design (25 points)

Include:

Examples:

Tips: Give higher marks for teams that show understanding of layer design and reasoning, not just copying code from online examples.

4. Implementation Plan & Tools (25 points)

Include:

STAGE 2 – MODEL TRAINING & EVALUATION

Goal: Assess ability to build, train, tune, and evaluate neural network models effectively.

Criteria Description Performance Levels Max Points
1. Data Preparation & Preprocessing Performs data cleaning, normalization, and splitting into train/test sets appropriately. ☐ Missing or wrong (0–25)
☐ Basic processing (26–50)
☐ Well-executed preprocessing with examples (51–75)
☐ Smart feature engineering and rationale (76–100)
25
2. Model Construction & Training Builds correct architecture, selects appropriate optimizer, loss, batch size, learning rate, etc. ☐ Code errors or non-functional (0–25)
☐ Runs but not explained (26–50)
☐ Works well, includes logs/graphs (51–75)
☐ Thorough experimentation, multiple setups compared (76–100)
25
3. Model Evaluation Uses multiple metrics (accuracy, F1-score, ROC, confusion matrix, etc.) to assess performance. ☐ Accuracy only (0–25)
☐ 1–2 additional metrics (26–50)
☐ Full evaluation with interpretation (51–75)
☐ Comprehensive comparison and analysis (76–100)
25
4. Optimization & Fine-tuning Applies regularization, dropout, early stopping, hyperparameter tuning, etc. ☐ None (0–25)
☐ Minimal tuning (26–50)
☐ Multiple techniques applied (51–75)
☐ In-depth optimization with trade-off analysis (76–100)
25
    Stage 2 Total /100

1. Data Preparation & Preprocessing

Description:
Performs data cleaning, normalization, and splitting into train/test sets appropriately.

Guidelines:

Hints / Tips:


2. Model Construction & Training

Description:
Builds correct architecture, selects appropriate optimizer, loss function, batch size, and learning rate.

Guidelines:

Hints / Tips:


3. Model Evaluation

Description:
Uses multiple metrics (accuracy, F1-score, ROC, confusion matrix, etc.) to assess performance.

Guidelines:

Hints / Tips:


4. Optimization & Fine-tuning

Description:
Applies regularization, dropout, early stopping, hyperparameter tuning, etc.

Guidelines:

Hints / Tips:

STAGE 3 – APPLICATION & DEPLOYMENT

Goal: Evaluate ability to integrate, deploy, and present the trained model as a usable product.

Criteria Description Performance Levels Max Points
1. Model Application Integrates the model into a usable system (API, GUI, or product demo). ☐ Not applied (0–25)
☐ Simple local demo (26–50)
☐ Working application with clear inputs/outputs (51–75)
☐ Fully functional real-world application (76–100)
25
2. Deployment Deploys the model on a cloud or server (e.g., Flask, FastAPI, Streamlit, Docker). ☐ Not deployed (0–25)
☐ Manual/local deployment (26–50)
☐ Automated deployment with documentation (51–75)
☐ Full CI/CD or public web deployment (76–100)
25
3. Visualization & Results Presentation Includes visual dashboards, charts, or performance reports. ☐ Missing visuals (0–25)
☐ Basic plots only (26–50)
☐ Clear dashboard or graphs (51–75)
☐ Professional visualization and insights (76–100)
25
4. Final Report & Presentation Final report or slides explain pipeline, results, limitations, and future work. ☐ Incomplete (0–25)
☐ Adequate but lacks clarity (26–50)
☐ Well-structured report with visuals (51–75)
☐ Professional, logical, and insightful presentation (76–100)
25
    Stage 3 Total /100

1. Model Application

Description:
Integrates the trained model into a usable system (API, GUI, or product demo).

Guidelines:

Hints / Tips:


2. Deployment

Description:
Deploys the model on a server, cloud platform, or containerized environment.

Guidelines:

Hints / Tips:


3. Visualization & Results Presentation

Description:
Includes visual dashboards, charts, or performance reports for user understanding.

Guidelines:

Hints / Tips:


4. Final Report & Presentation

Description:
Provides a final report or slides explaining the pipeline, results, limitations, and future work.

Guidelines:

Hints / Tips: