A Python package for analyzing multilingual text.
multilang-probe is a toolkit designed to classify character sets, detect languages in text files, and extract specific multilingual passages. It supports character detection for a wide range of writing systems using Unicode script properties (e.g., Latin, Japanese, Cyrillic, Arabic, Devanagari, and more). Additionally, it leverages the FastText model for robust language detection.
Whether you are analyzing large corpora or extracting specific language data, multilang-probe simplifies the process with an easy-to-use API.
- Detect and calculate proportions of character types (e.g., Latin, Japanese, Cyrillic, Arabic, Devanagari) in text.
- Uses
regex
with Unicode script properties (\p{Script}
) for more accurate classification. - Special handling for Japanese vs Chinese characters (Han script).
from multilang_probe.character_detection import classify_text_with_proportions
# Sample text with multiple languages/scripts
text = "これは日本語です。Привет мир! Ελληνικά και हिन्दी।"
# Classify the text
proportions = classify_text_with_proportions(text)
# Print the proportions
print("Character script proportions:")
print(proportions)
Expected outcome:
Character script proportions:
{'japanese': 19.51, 'cyrillic': 21.95, 'greek': 26.83, 'devanagari': 14.63, 'other': 17.07}
Explanation:
- If the text contains Hiragana/Katakana, Han characters are considered Japanese Kanji.
- Otherwise, Han characters are considered Chinese.
- Identify top languages in text using Facebook's FastText pre-trained model.
from charlang_detect.language_detection import detect_language_fasttext
text = "Ceci est un texte en français."
languages = detect_language_fasttext(text)
print(languages)
# Output example: "fr: 99.2%, en: 0.8%"
- Analyze all
.txt
files in a folder to detect multilingual passages and language distributions. - Character-based filtering: Identify and filter text lines containing specific character sets (e.g., Japanese, Cyrillic, Arabic).
- Language-based filtering: Extract passages in a specific language, with customizable confidence thresholds (e.g., 70%).
- Targeted extraction: Extract lines of text meeting both minimum length requirements and language detection accuracy.
- Calculate language proportions: Aggregate detected languages across files and calculate their proportions.
from charlang_detect.corpus_analysis import analyze_corpus_with_fasttext
folder_path = "path/to/corpus/"
results = analyze_corpus_with_fasttext(folder_path)
for filename, langs in results.items():
print(filename, langs)
from charlang_detect.corpus_analysis import filter_passages_by_language
folder_path = "path/to/corpus/"
target_languages = ["fr", "en"]
threshold = 70
filtered = filter_passages_by_language(results, target_languages, folder_path, threshold)
for filename, passages in filtered.items():
print(filename, passages)
- Japanese (Hiragana, Katakana)
- Han (Kanji; considered Japanese if Hiragana/Katakana present, else Chinese)
- Korean (Hangul)
- Cyrillic (for languages like Russian, Bulgarian, etc.)
- Arabic
- Hebrew
- Greek
- Latin (basic and extended)
- Devanagari (e.g., Hindi, Sanskrit)
- Tamil, Bengali, Thai
- Extendable via Unicode scripts
- "other" category for characters not belonging to known scripts
- Python 3.7+
- FastText
- Regex (for Unicode script classification)
This project is licensed under the MIT License. While the MIT License allows unrestricted use, modification, and distribution of this software, I kindly request that proper credit be given when this project is used in academic, research, or published work. For citation purposes, please refer to the following:
CAFIERO Florian, 'multilang-probe', 2024, [https://github.com/floriancafiero/multilang-probe].
Pull requests are welcome! For major changes, please open an issue first to discuss what you would like to change.
Florian Cafiero
GitHub: floriancafiero
Email: [email protected]
- Support for other pre-trained language models (e.g., spaCy).
- Detection of mathematical language
- Visualization tools for multilingual analysis.
- CLI (Command-Line Interface) for easy usage without writing code.