Skip to content

Commit

Permalink
Merge pull request #20 from SF-Nexus/hawc2-patch-4
Browse files Browse the repository at this point in the history
Update _index.md
  • Loading branch information
hawc2 authored Jan 23, 2024
2 parents 0f3f59b + bdab382 commit 0a4fee6
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions content/Data/_index.md
Original file line number Diff line number Diff line change
Expand Up @@ -35,11 +35,11 @@ from datasets import load_dataset
dataset = load_dataset("SF-Corpus/EF_Chapters_and_Chunks")
```

For more information about working with HuggingFace datasets, review their reference guide: https://huggingface.co/docs/datasets/v1.1.1/loading_datasets.html
For more information about working with HuggingFace datasets, review their [reference guide](https://huggingface.co/docs/datasets/v1.1.1/loading_datasets.html)

An extended discussion on extracted features can be found on the Scholars' Studio blog: https://sites.temple.edu/tudsc/2019/07/18/curating-copyrighted-corpora-an-r-script-for-extracting-and-structuring-textual-features/
An extended discussion on extracted features can be found on the Scholars' Studio blog, in Jeff Antsen's ["Curating Copyrighted Corpora: AN R Script for Extracting and Structuring Textual Features"](https://sites.temple.edu/tudsc/2019/07/18/curating-copyrighted-corpora-an-r-script-for-extracting-and-structuring-textual-features/).

This project made use of multiple Python and R pipelines to extract features from the science fiction collection. These pipelines are available as both Jupyter Notebooks and Google Colab Notebooks in this Github repository: https://github.com/SF-Nexus/extracted-features/tree/main/notebooks. Below, the process for crafting each extracted features dataset is discussed in more detail.
This project made use of multiple Python and R pipelines to extract features from the science fiction collection. These pipelines are available as both Jupyter Notebooks and Google Colab Notebooks in this [Github repository](https://github.com/SF-Nexus/extracted-features/tree/main/notebooks). Below, the process for crafting each extracted features dataset is discussed in more detail.

## Pipeline 1: Text Sectioning and Disaggregation
*Full Code Available on SF Nexus Github:*
Expand Down

0 comments on commit 0a4fee6

Please sign in to comment.