Skip to content

benymng/hack-mit

Repository files navigation

Getting Started

First, run the development server:

npm run dev
# or
yarn dev
# or
pnpm dev

Open http://localhost:3000 with your browser to see the result.

You can start editing the page by modifying app/page.js. The page auto-updates as you edit the file.

This project uses next/font to automatically optimize and load Inter, a custom Google Font.

Learn More

To learn more about Next.js, take a look at the following resources:

You can check out the Next.js GitHub repository - your feedback and contributions are welcome!

Deploy on Vercel

The easiest way to deploy your Next.js app is to use the Vercel Platform from the creators of Next.js.

Check out our Next.js deployment documentation for more details.

Table of Contents

Introduction

This project is focused on document scanning and health data analysis using computer vision and machine learning. It combines various technologies to provide an innovative solution for document recognition, health data prediction, and community matching. This README provides an overview of the project and its components.

Project Overview

The project comprises both frontend and backend components:

  • Frontend:

    • Sign-up / Sign-in Page
    • Home Page
    • Data Display Page
    • Community Page
    • Integration of JavaScript Mobile Document Scanner
    • Integration of user data storage
  • Backend:

    • Processing text from images and associating key-value pairs
    • Utilizes Django as the web framework
    • Setting up vector databases and matching
    • Creating a method of communication within a group/community
    • Data communication between frontend and backend
    • Neural Network models for prediction

The core process involves taking pictures of paper documents, extracting key information using computer vision, user questionnaires processed through LLMs, and matching users to unique communities based on health data and other factors.

Technology Stack

The project utilizes a diverse technology stack:

  • File Managment

    • Github
    • Git
    • Git Reposetory
  • Frontend:

    • Next.js
    • JavaScript
    • React
    • Tailwind CSS
  • Backend:

    • Firebase
    • PineconeDB
  • Computer Vision:

    • TensorFlow
    • Tesseract
  • Presentation:

    • Canva
    • Figma
    • Google Slides

Team and Responsibilities

  • John: Figma Mockup, Frontend Application of Mobile Site, Pitch Deck
  • Siya: Predictive Models, Computer Vision, Frontend, Pitch Deck
  • Ben: Computer Vision, Predictive Models, Vector Database, Github & Git Reposetory Setup
  • Auston: Backend (API), Computer Vision, Messaging Integration

Development Timeline

Here's an overview of the project development timeline:

  • Integration and pitch deck: 6 hours
  • Figma design: 2.5 hours
  • Frontend development: 12 hours
  • Computer vision MVP: 12 hours
  • Predictive model MVP: 8 hours
  • Community matching via vector database: 6 hours

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •