Python library for computing propositional idea density.
- Introduction
- What is Idea Density?
- Installation
- Usage
- Requirements
- Development Setup
- Running Tests
- CPIDR Parity with CPIDR 3.2
- References
- Citing
- Contributing
- License
ideadensity is a Python library which determines the propositional idea density of an English text automatically. This project aims to make this functionality more accessible to Python developers and researchers. ideadensity provides two ways of computing idea density:
- CPIDR. The CPIDR implementation in ideadensity is a direct port of the Computerized Propositional Idea Density Rater (CPIDR) 3.2 (Brown et al., 2008) [1]
- DEPID. This library implements the DEPID algorithm described by Sirts et al (2017) [2]
Here's a quick example of how to use ideadensity:
from ideadensity import cpidr, depid
text = "The quick brown fox jumps over the lazy dog."
cpidr_word_count, proposition_count, cpidr_density, word_list = cpidr(text)
depid_density, depid_word_count, dependencies = depid(text)
print(f"CPIDR density: {cpidr_density:.3f}")
print(f"DEPID density: {depid_density:.3f}")
Idea density, also known as propositional density, is a measure of the amount of information conveyed relative to the number of words used. It's calculated by dividing the number of expressed propositions by the number of words. This metric has applications in various fields, including linguistics, cognitive science, and healthcare research.
- Install the package
pip install ideadensity
- Download the required spaCy model:
python -m spacy download en_core_web_sm
poetry add ideadensity
python -m spacy download en_core_web_sm
Note: This package currently supports Python 3.10-3.12 due to dependency constraints with spaCy and its dependencies. If you're using Python 3.13, you'll need to create a virtual environment with a compatible Python version.
Here's a simple example of how to use CPIDR:
from ideadensity import cpidr
text = "The quick brown fox jumps over the lazy dog."
word_count, proposition_count, density, word_list = cpidr(text)
print(f"Word count: {word_count}")
print(f"Proposition count: {proposition_count}")
print(f"Idea density: {density:.3f}")
# Analyzing speech
speech_text = "Um, you know, I think that, like, the weather is nice today."
word_count, proposition_count, density, word_list = cpidr(speech_text, speech_mode=True)
print(f"Speech mode - Idea density: {density:.3f}")
# Detailed word analysis
for word in word_list.items:
if word.is_word:
print(f"Token: {word.token}, Tag: {word.tag}, Is proposition: {word.is_proposition}")
ideadensity CPIDR mode supports a speech mode that handles common speech patterns and fillers differently from written text. When analyzing transcripts or spoken language, use the speech_mode=True
parameter for more accurate results.
Here's an example of how to use the DEPID functionality:
from ideadensity import depid
text = "The quick brown fox jumps over the lazy dog."
density, word_count, dependencies = depid(text)
print(f"Word count: {word_count}")
print(f"Idea density: {density:.3f}")
print("Dependencies:")
for dep in dependencies:
print(f"Token: {dep[0]}, Dependency: {dep[1]}, Head: {dep[2]}")
DEPID-R counts distinct dependencies.
from ideadensity import depid
text = "This is a test of DEPID-R. This is a test of DEPID-R"
density, word_count, dependencies = depid(text, is_depid_r=True)
print(f"DEPID-R idea density: {density:.3f}")
ideadensity DEPID mode supports custom filtering of sentences and tokens. By default, ideadensity uses filters described by (Sirts et al., 2017):
- Sentence filter.
- Filter out sentences with "I" or "You" as the subject of the sentence (i.e. if the "I" or "You" token dependency is "nsubj" and it's head dependency is the root).
- Note: Sirts et al (2017) also filters out vague sentences using SpeciTeller. That is a filter which ideadensity does not yet implement.
- Token filters:
- Filter out "det" dependencies if the token is "a", "an" or "the".
- Filter out "nsubj" dependencies if the token is "it" or "this".
- Filter out all "cc" dependencies.
This example demonstrates how to apply your own custom filters to modify the analysis. The sentence_filters
and token_filters
parameters allow you to customize the DEPID algorithm to suit your specific needs.
def custom_sentence_filter(sent):
return len(sent) > 3
def custom_token_filter(token):
return token.pos_ != "DET"
text_with_filters = "I run. The quick brown fox jumps over the lazy dog."
density, word_count, dependencies = depid(text_with_filters,
sentence_filters=[custom_sentence_filter],
token_filters=[custom_token_filter])
print(f"\nWith custom filters - Idea density: {density:.3f}")
The package includes a command line interface for quick analysis of text: Command line options:
--text TEXT
: Directly provide text for analysis (can include multiple words)--file FILE
: Path to a file containing text to analyze--speech-mode
: Enable speech mode for analyzing transcripts (filters common fillers)--csv CSV
: Export token details to a CSV file at the specified path--txt TXT
: Export results to a TXT file in CPIDR format at the specified path
Note: You must provide either --text
or --file
when using the command line interface.
# Analyze text directly from command line
python main.py --text "The quick brown fox jumps over the lazy dog."
# Analyze text from a file
python main.py --file sample.txt
# Use speech mode with text from a file
python main.py --file transcript.txt --speech-mode
# Export token details to a CSV file
python main.py --text "This is a test sentence." --csv output.csv
# Export results in CPIDR-compatible format to a TXT file
python main.py --text "This is a test sentence." --txt output.txt
# Export in both formats
python main.py --file sample.txt --csv output.csv --txt output.txt
Use one of the provided downloads for your operating system, or clone this repository and run:
python main.py
CSV Export: The CSV export includes detailed information about each token with the following columns:
- Token: The actual word or token
- Tag: The part-of-speech tag
- Is Word: Whether the token is considered a word (True/False)
- Is Proposition: Whether the token is considered a proposition (True/False)
- Rule Number: The rule number that identified the token as a proposition (if applicable)
TXT Export: The TXT export produces a file in a format compatible with the original CPIDR tool:
ideadensity 0.2.11
"This is a test sentence...."
054 PRP W This
200 VBZ W P is
201 DT W a
200 JJ W P test
NN W sentence
. .
2 propositions
5 words
0.400 density
Each line in the token section includes:
- Rule number (if available)
- Part-of-speech tag
- Word marker (W if the token is a word)
- Proposition marker (P if the token is a proposition)
- The token text
- Python 3.10+
- spaCy 3.7.5+
To set up the development environment:
- Clone the repository
- Install Poetry if you haven't already:
pip install poetry
- Install project dependencies:
poetry install
- Install the required spaCy model:
poetry run python -m spacy download en_core_web_sm
- Activate the virtual environment:
poetry shell
To run the tests, use pytest:
pytest tests/
Because this port uses spaCy as a part-of-speech tagger instead of the original program's MontyLingua, there is a very slight difference in the reported idea density. This port includes unit tests containing 847 words of text. ideadensity: 434 propositions. 0.512 idea density CPIDR 3.2: 436 propositions. 0.515 idea density
For more information about the original CPIDR 3.2, please visit CASPR's official page.
[1] Brown, C., Snodgrass, T., Kemper, S. J., Herman, R., & Covington, M. A. (2008). Automatic measurement of propositional idea density from part-of-speech tagging. Behavior research methods, 40(2), 540-545.
[2] Sirts, K., Piguet, O., & Johnson, M. (2017). Idea density for predicting Alzheimer's disease from transcribed speech. arXiv preprint arXiv:1706.04473.
If you use this project in your research, you may cite it as:
Jason Robison. (2024). ideadensity (0.2.0) [Source code]. GitHub. https://github.com/jrrobison1/ideadensity
Contributions are welcome! Please feel free to submit a Pull Request.
- Fork the repository
- Create your feature branch (
git checkout -b feature/AmazingFeature
) - Commit your changes (
git commit -m 'Add some AmazingFeature'
) - Push to the branch (
git push origin feature/AmazingFeature
) - Open a Pull Request
Please ensure that your code passes all tests and follows the project's coding style.
This project is licensed under the GNU General Public License v2.0. See the LICENSE file for details.
ideadensity's CPIDR implementation is a port of the original CPIDR 3.2, which was released under GPL v2. This project maintains the same license to comply with the terms of the original software.