Skip to content

suhasml/Lip-Reader

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Lip Reader

This lip reader project utilizes a complex neural network to generate captions for videos based on lip movements. The trained model can infer and comprehend spoken language from lip movements in video frames. The lip reading system has been trained on a substantial dataset of videos with corresponding captions, enabling it to recognize and interpret lip movements accurately.

Requirements

  • Python 3.x
  • TensorFlow (version 2.x)
  • OpenCV
  • NumPy
  • Pandas

Usage

To use the lip reader model for generating captions on new videos:

  1. Clone this repository to your local machine.
  2. Prepare your video files and corresponding alignments (captions).
  3. Customize the load_video() and load_alignments() functions in the load_data() method to read your data correctly.
  4. Update the path to your trained model in the predict() method.
  5. Utilize the provided code to preprocess video frames, perform inference, and generate captions.

Contribution

Contributions to this project are welcome! If you encounter any issues or have suggestions for improvements, please feel free to open an issue or submit a pull request.

License

This project is licensed under the MIT License - see the LICENSE file for details.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published