Auto cut videos by silences
This project is a comprehensive command-line toolkit designed for video editing and processing tasks. It provides a modular framework to perform operations such as trimming by silence, adding subtitles, denoising audio, generating transcripts and translations, and even creating short videos with dynamic titles.
config.json
) to configure various operations.pip
):moviepy
numpy
scipy
whisper
torch
torchaudio
pydub
librosa
bark
(for audio generation)Clone the repository:
git clone <repository_url>
cd <repository_directory>
Create and activate a virtual environment (recommended):
python -m venv venv
source venv/bin/activate # On Windows: venv\Scripts\activate
Install the required packages:
pip install -r requirements.txt
(Ensure your requirements.txt
lists all necessary dependencies.)
Configure the project by editing the config.json
file to adjust settings such as subtitle and title clip configurations.
The main entry point is main.py
, which provides several subcommands for different tasks. Run the following command to see available options:
python main.py --help
The CLI tool uses subcommands to select the desired functionality. You invoke the tool using one of the following subcommands:
Each subcommand has its own required and optional arguments. Use the --help
flag with any subcommand to see detailed usage information.
video_edit
)Description:
This subcommand processes video files by applying a sequence (pipeline) of editing functions.
python main.py video_edit <input_file1> [<input_file2> ...] --pipeline <function1> [<function2> ...] [options]
input_file (required):
One or more video file paths to process.
—pipeline (required):
List of functions to apply to each video file.
Available functions:
trim_by_silence
denoise
transcript
subtitles
save_separated_video
save_join
save_video
set_vertical
set_horizontal
-c, —clip_interval:
Type: float, Default: 2
Description: Clipping precision.
-s, —sound_threshold:
Type: float, Default: 0.01
Description: Maximum volume threshold to consider silence.
-d, —discard_silence:
Type: boolean flag (uses a string-to-boolean converter), Default: False
Description: Discard silent clips.
python main.py video_edit video1.mp4 video2.mp4 --pipeline trim_by_silence subtitles -c 3 -s 0.02 -d True
This applies the trim_by_silence
and subtitles
functions to video1.mp4
and video2.mp4
with a clip interval of 3 seconds and a sound threshold of 0.02, discarding silent clips.
separate_audio
)Description:
Extracts audio from the given video files.
python main.py separate_audio <video_file1> [<video_file2> ...]
python main.py separate_audio video1.mp4 video2.mp4
This command will extract the audio from video1.mp4
and video2.mp4
and save them accordingly.
split_str
)Description:
Splits SRT subtitle files into smaller segments based on a specified number of words per subtitle.
python main.py split_str <srt_file1> [<srt_file2> ...] <words_per_subtitle>
files (required):
One or more SRT files to split.
words_per_subtitle (required):
Type: integer
Description: The number of words per subtitle segment.
python main.py split_str subtitles.srt 5
This will split the subtitles in subtitles.srt
so that each subtitle contains approximately 5 words.
voice
)Description:
Performs voice operations such as video translation or audio generation.
python main.py voice <operation> <video_path> [options]
operation (required):
Operation to perform. Choose between:
video_translation
audio_generator
video_path (required):
The path to the video file to process.
-t, —translate:
Type: Model to use for translation, Default: "Helsinki-NLP/opus-mt-es-en"
Description: Translate the video to English or transcribe in the same language.
—voice:
Type: string, Default: "en-us/af_heart"
Description: Voice model to use for translation.
python main.py voice video_translation video1.mp4 -t Helsinki-NLP/opus-mt-es-en --voice en-us/af_heart
This translates video1.mp4
using the specified voice model, with translation turned off (if you only want transcription).
generator
)Description:
Generates a base video or adds titles to a short video.
python main.py generator <file1> [<file2> ...] <tool>
files (required):
One or more files to process.
tool (required):
The tool to use. Available options:
base
– to generate a base video.add_titles
– to add titles to the video.
python main.py generator video1.mp4 base
This command uses the base
tool on video1.mp4
to generate a base video.
To display the help information for the CLI tool or a specific subcommand, use the --help
flag. For example:
python main.py --help
python main.py video_edit --help
This will display all available options and arguments for that command.
config.json
and makes it available throughout the project.The toolkit uses a JSON configuration file (config.json
) to define parameters such as:
Adjust these settings according to your needs before running any commands.
Contributions are welcome! If you have suggestions or improvements, feel free to open an issue or submit a pull request.
This project is licensed under the MIT License. See the LICENSE file for details.