From 4c593ed50b31f8d2640c89114ec826a5d4ca7649 Mon Sep 17 00:00:00 2001 From: Hammouda Elbez <hammouda.elbez@univ-lille.fr> Date: Thu, 3 Nov 2022 10:37:01 +0100 Subject: [PATCH] Requirements updated --- README.md | 3 ++- requirements.txt | 26 +++++++++++++------------- src/Modules/General/spark.py | 1 - src/Modules/Synapse/spark.py | 1 - 4 files changed, 15 insertions(+), 16 deletions(-) diff --git a/README.md b/README.md index 9da7721..14d6187 100755 --- a/README.md +++ b/README.md @@ -51,7 +51,7 @@ pip install -r requirements.txt - Add MongoDB credentials in `config.py` -4. Using Python >= 3.0 (**make sure MongoDB is running**) run `python VS2N.py` to start VS2N. If everything is fine, the web interface should appear automatically. +4. Using Python >= 3.0 (**make sure MongoDB is running**), run `python VS2N.py` to start VS2N. If everything is fine, the web interface should appear automatically. ## Docker image You can also use a docker image to run VS2N : https://hub.docker.com/r/helbez/vs2n @@ -79,6 +79,7 @@ src/ # Source directory. ### Core - Multilayer support added - Spark & MongoDB version updated +- Added support for more simulators (Norse `Python`, CSNN `C++`) ## v0.35 diff --git a/requirements.txt b/requirements.txt index c4bf4df..618cc18 100755 --- a/requirements.txt +++ b/requirements.txt @@ -1,16 +1,16 @@ pypandoc -pymongo==4.1.0 -pyspark==3.2.1 +pymongo +pyspark pandas -dash==2.3.1 -plotly==5.7.0 -dash-renderer==1.9.1 -dash-html-components==2.0.0 -dash-core-components==2.0.0 -dash-bootstrap-components==1.1.0 -dash-cytoscape==0.3.0 -dash-daq==0.5.0 -Flask-Login==0.6.0 -wheel==0.37.1 -werkzeug==2.1.1 +dash +plotly +dash-renderer +dash-html-components +dash-core-components +dash-bootstrap-components +dash-cytoscape +dash-daq +Flask-Login +wheel +werkzeug PyArrow diff --git a/src/Modules/General/spark.py b/src/Modules/General/spark.py index c80ad8a..697516d 100755 --- a/src/Modules/General/spark.py +++ b/src/Modules/General/spark.py @@ -1,7 +1,6 @@ """ Spark pre-processing operations. """ -from operator import contains from bson.json_util import dumps from bson.json_util import loads import traceback diff --git a/src/Modules/Synapse/spark.py b/src/Modules/Synapse/spark.py index 68db8c7..d577247 100755 --- a/src/Modules/Synapse/spark.py +++ b/src/Modules/Synapse/spark.py @@ -4,7 +4,6 @@ import pandas as pd import pymongo import traceback -from pyspark.sql import functions as F from src.templates.sparkOp import sparkOp def init(g): -- GitLab