Skip to content

hhucn/Example-Implementation-2

Repository files navigation

Example Recommendation System 2

Here, we see an example implementation of the comment recommendation framework.

Model

This model recommends comments based on their popularity. By this, the user should get an overview of the most popular opinions about the topic of the discussion from different sources and articles.

For this, the model searches for articles that are similar to the article the comment appeared under the user is interested in. It takes all comments published under these articles and sorts them accordingly to the number of up-votes they received.

Framework

The Comment Recommendation Framework is a modular approach to support scientists in the development of prototypes for comment recommendation systems that can be used in real-world scenarios. The advantage of such a system is that it relieves the scientist from the majority of the technical code and only prototype-specific components have to be developed. In this way, the researchers can invest more time in the development of recommendation models and less time has to be spent in the development of a prototype while at the same time getting prototypes that can be used in real-world settings.

Implementation Effort

To implement this example recommendation system. The following files have been edited:

  • DB/db_models/comment.py -> Here, we added the property up_votes.
  • Embedder/embedding_model.py -> Here, we implemented the method to compute the vector embeddings.
  • Embedder/run_embedder.py -> Here, we determined for which properties the embeddings are computed.
  • Model/model.py -> Here, we implemented the recommendation model.
  • NewsAgencyScraper/NewsAgencyScraper/spiders/WashingtonTimesSpyder.py -> Here, we replaced the template class NewsAgencySpyder with the implementation for the WashingtonTimesSpyder.
  • NewsAgencyScraper/NewsAgencyScraper/spiders/NewYorkTimesSpyder.py -> Here, we replaced the template class NewsAgencySpyder with the implementation for the NewYorkTimesSpyder.
  • NewsAgencyScaper/NewsAgencyScraper/pipelines.py -> Here, we updated the method process_item to store the new property up_votes in the database.
  • NewsAgencyScraper/run_scraper.py -> Here, we called the WashingtonTimesSpyder and NewYorkTimesSpyder.
  • UI folder -> Here, we installed the npm packages and built the chrome extension

Setup

Ensure that the following tools are installed:

  • Docker
  • Docker-Compose
  • Python >= 3.10

Documentation

To build the latest version of the documentation, please run in the docs folder:

$ make clean && make html

Then you find the latest documentation here

Environment Variables

The framework need some environment variables to be set for running properly. Please ensure that you have an .env file with the following variables:

  • NEO4J_PASSWORD
  • NEO4J_BOLT_URL (Format: bolt://neo4j:<NEO4J_PASSWORD>@neo4j:7687)

Run different moduls with docker-compose

We provide you with the following docker-compose files to run the different components of the example implementation.

  • docker-compose.scraping.yml: Runs the news agency scraper to retrieve articles and comments from various news agencies.
  • docker-compose.embed.yml: Starts the embedding process to compute the embeddings for the comments and articles. Should be run directly after docker-compose.scraping.yml.
  • docker-compose.test.yml: Runs the tests for the system.
  • docker-compose.api.yml: Runs the comment-recommendation systems.

User-Interface

If you would like to use the carousel view user-interface, you have to install the npm packages and build the chrome extension. For this you have to run in the UI folder:

$ npm install

and afterwards:

$ npm run build

Then you can import the build folder in a chromium browser.

Maintainers:

  • Anonymous

Contributors:

  • Anonymous

License:

Copyright(c) 2022 - today Anonymous

Distributed under the MIT License

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published