jonathanagustin commited on
Commit
a3c770d
1 Parent(s): 7dbaf21

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -26
README.md CHANGED
@@ -4,7 +4,7 @@ emoji: 🎥
4
  colorFrom: blue
5
  colorTo: green
6
  sdk: gradio
7
- python_version: '3.8'
8
  sdk_version: 4.44.0
9
  app_file: app.py
10
  tags:
@@ -20,56 +20,75 @@ datasets:
20
  ---
21
  # Object Detection in Live YouTube Streams
22
 
23
- ## Project Status: Active
24
-
25
  ## Installation
26
 
27
  To use and install this project, follow these steps:
28
 
29
- 1. Clone the repository from GitHub.
30
- 2. Ensure Python 3.8 or higher is installed on your machine.
31
- 3. Install required dependencies using `pip install -r requirements.txt`.
32
- 4. Run `python3 app.py` file to start the application.
 
 
 
 
 
 
 
 
 
 
 
 
 
33
 
34
  ## Objective
35
 
36
  The primary goal of this project is to harness computer vision and machine learning technologies for real-time, accurate object detection in live YouTube streams. By focusing on this, we aim to unlock new potential in areas critical to modern society, such as enhanced security surveillance, efficient urban management, and advanced traffic analysis systems. Our objective is to develop a robust system that not only identifies and classifies objects in diverse streaming environments but also adapts to varying conditions with high precision and reliability.
37
 
38
- ## Contributors
39
-
40
- - **William Acuna**
41
- - **Jonathan Agustin**
42
- - **Alec Anderson**
43
-
44
  ## Methods Used
45
 
46
- - **Computer Vision and Object Detection**: Computer vision techniques and object detection models identified and classified objects in live video feeds.
47
- - **Machine Learning and Deep Learning**: Machine learning, especially deep learning, interpreted complex visual data from video streams.
48
- - **Data Streaming**: Efficient data streaming methods handled the live video feeds from online sources.
49
- - **User Interface Design**: A user-friendly interface enabled simple interaction with the system, including video input and result visualization.
50
- - **API Integration for Video Retrieval**: API solutions retrieved of live video content from popular online platforms.
 
51
 
52
  ## Technologies
53
 
54
  - **Python**: Primary programming language for the project's development.
55
  - **Git**: Version Control System for tracking and managing changes in the codebase.
56
  - **GitHub**: Platform for code hosting, collaboration, and version control.
57
- - **YouTube API**: Data source for accessing live YouTube streams.
58
- - **Ultralytics (YOLOv8)**: Object detection model for real-time video analysis.
59
- - **Google Colab**: Cloud-based platform for development and testing of the model.
60
- - **Hugging Face Spaces**: Deployment service for hosting the machine learning model.
61
- - **Gradio**: Framework for building the user interface and facilitating user interactions.
 
 
62
 
63
  ## Project Description
64
 
65
- This project is centered around the creation and deployment of a sophisticated object detection system, specifically tailored for live YouTube streams. Utilizing the advanced capabilities of the Ultralytics YOLO model, this system is designed to identify, classify, and track objects in real-time within dynamic streaming environments. A key aspect of our endeavor is to address and overcome challenges associated with variable lighting conditions, object occlusions, and diverse environmental settings, ensuring the system's effectiveness and accuracy in real-world applications. Moreover, we aim to optimize the system for speed and efficiency, ensuring minimal latency in real-time processing. The project not only represents a significant advancement in computer vision but also offers a versatile tool with wide-ranging applications, from urban planning and public safety to traffic management and surveillance.
 
 
 
 
 
 
 
 
 
 
 
66
 
67
  ## License
68
 
69
- This project is licensed under the MIT License - see the LICENSE.md file for details.
70
 
71
  ## Acknowledgments
72
 
73
  - **Professor Roozbeh Sadeghian**: Our advisor, for invaluable guidance and mentorship.
74
  - **Professor Ebrahim Tarshizi**: The Academic Director for the Applied Artificial Intelligence (AAI) program, for contributions to program structure and academic enrichment.
75
- - **The Applied Artificial Intelligence Program at the University of San Diego**: For essential support and resources.
 
4
  colorFrom: blue
5
  colorTo: green
6
  sdk: gradio
7
+ python_version: '3.10'
8
  sdk_version: 4.44.0
9
  app_file: app.py
10
  tags:
 
20
  ---
21
  # Object Detection in Live YouTube Streams
22
 
 
 
23
  ## Installation
24
 
25
  To use and install this project, follow these steps:
26
 
27
+ 1. **Clone the repository from GitHub**:
28
+ ```bash
29
+ git clone https://huggingface.co/spaces/aai521-group6/youtube-object-detection
30
+ ```
31
+ 2. **Navigate to the project directory**:
32
+ ```bash
33
+ cd youtube-object-detection
34
+ ```
35
+ 3. **Ensure Python 3.8 or higher is installed on your machine**.
36
+ 4. **Install required dependencies using**:
37
+ ```bash
38
+ pip install -r requirements.txt
39
+ ```
40
+ 5. **Run the application**:
41
+ ```bash
42
+ python app.py
43
+ ```
44
 
45
  ## Objective
46
 
47
  The primary goal of this project is to harness computer vision and machine learning technologies for real-time, accurate object detection in live YouTube streams. By focusing on this, we aim to unlock new potential in areas critical to modern society, such as enhanced security surveillance, efficient urban management, and advanced traffic analysis systems. Our objective is to develop a robust system that not only identifies and classifies objects in diverse streaming environments but also adapts to varying conditions with high precision and reliability.
48
 
 
 
 
 
 
 
49
  ## Methods Used
50
 
51
+ - **Computer Vision and Object Detection**: Implemented advanced computer vision techniques and object detection models to identify and classify objects in live video feeds.
52
+ - **Machine Learning and Deep Learning**: Leveraged machine learning, especially deep learning, to interpret complex visual data from video streams.
53
+ - **Asynchronous Processing**: Integrated asynchronous processing to improve the performance and responsiveness of the application.
54
+ - **Data Streaming**: Employed efficient data streaming methods to handle live video feeds from online sources.
55
+ - **User Interface Design**: Designed an enhanced, user-friendly interface with Gradio, enabling simple interaction with the system, including video input and result visualization.
56
+ - **API Integration for Video Retrieval**: Utilized API solutions, such as `youtube-search-python` and Streamlink, to retrieve live video content from popular online platforms.
57
 
58
  ## Technologies
59
 
60
  - **Python**: Primary programming language for the project's development.
61
  - **Git**: Version Control System for tracking and managing changes in the codebase.
62
  - **GitHub**: Platform for code hosting, collaboration, and version control.
63
+ - **YouTube API and Libraries**: Data source for accessing live YouTube streams, using libraries like `youtube-search-python`.
64
+ - **Ultralytics YOLOv8**: Object detection model for real-time video analysis.
65
+ - **OpenCV**: Library for image and video processing tasks.
66
+ - **Gradio**: Framework for building an interactive and user-friendly interface.
67
+ - **Streamlink**: Tool for extracting live stream URLs.
68
+ - **Imageio**: Used for reading frames from live streams using FFmpeg.
69
+ - **Asyncio**: Enables asynchronous processing to improve application performance.
70
 
71
  ## Project Description
72
 
73
+ This project is centered around the creation and deployment of a sophisticated object detection system, specifically tailored for live YouTube streams. Utilizing the advanced capabilities of the Ultralytics YOLOv8 model, this system is designed to identify, classify, and track objects in real-time within dynamic streaming environments.
74
+
75
+ Recent updates to the project include:
76
+
77
+ - **Comprehensive Code Refactoring**: Improved efficiency and maintainability of the codebase by restructuring and optimizing code.
78
+ - **Maintained Docstring Style**: Ensured consistent and detailed documentation throughout the code for better readability and understanding.
79
+ - **Enhanced User Interface and Experience**: The user interface has undergone a significant makeover using Gradio, offering a more intuitive and engaging experience with modern theming, improved layout, and clear instructions.
80
+ - **Asynchronous Processing**: Implemented asynchronous functions where applicable using `asyncio`, enhancing the performance and responsiveness of the application, especially during network operations and long-running tasks.
81
+
82
+ A key aspect of our endeavor is to address and overcome challenges associated with variable lighting conditions, object occlusions, and diverse environmental settings, ensuring the system's effectiveness and accuracy in real-world applications. Moreover, we aim to optimize the system for speed and efficiency, ensuring minimal latency in real-time processing.
83
+
84
+ The project not only represents a significant advancement in computer vision but also offers a versatile tool with wide-ranging applications, from urban planning and public safety to traffic management and surveillance.
85
 
86
  ## License
87
 
88
+ This project is licensed under the MIT License - see the [LICENSE.md](LICENSE.md) file for details.
89
 
90
  ## Acknowledgments
91
 
92
  - **Professor Roozbeh Sadeghian**: Our advisor, for invaluable guidance and mentorship.
93
  - **Professor Ebrahim Tarshizi**: The Academic Director for the Applied Artificial Intelligence (AAI) program, for contributions to program structure and academic enrichment.
94
+ - **The Applied Artificial Intelligence Program at the University of San Diego**: For essential support and resources.