File tree Expand file tree Collapse file tree 3 files changed +14
-3
lines changed
frontend/src/pages/tagging
tagging-service/flaskr/endpoints Expand file tree Collapse file tree 3 files changed +14
-3
lines changed Original file line number Diff line number Diff line change @@ -35,8 +35,15 @@ In case there are issues due to dependencies try to rebuild the containers (will
35
35
docker-compose up --build # rebuild
36
36
```
37
37
38
- Note: The clustering might take a while, the clusters tend to finish close to each other don't panic if you see no
39
- progress for several minutes.
38
+ Notes:
39
+
40
+ - The clustering might take a while, the clusters tend to finish close to each other don't panic if you see no progress
41
+ for several minutes.
42
+ - If the tagging-service is interrupted before completing the clustering it'll be necessary to manually log into the
43
+ database and delete the object under ` dataset_db/dataset ` with ` dataset_id ` equal to the one that was interrupted.
44
+
45
+ If the entry is not deleted it'll be impossible to complete the clustering for it and the dataset will always result
46
+ as loading.
40
47
41
48
Optimizations: It's possible to change the resources allocated to the python-service from ` .env ` , this will impact
42
49
clustering performance, to change the resources modify the environment variables:
Original file line number Diff line number Diff line change @@ -59,7 +59,7 @@ function DatasetSelection() {
59
59
< TableBody >
60
60
{ datasets . map ( ( dataset : DatasetDesc ) => {
61
61
const loading_cluster = dataset . clusters_computed != dataset . nr_questions
62
- const needed_time_s = 1000 * 60 * 2 * dataset . nr_questions
62
+ const needed_time_s = 1000 * 90 * dataset . nr_questions
63
63
const started = new Date ( dataset . creation_data )
64
64
const now = new Date ( )
65
65
Original file line number Diff line number Diff line change @@ -66,6 +66,10 @@ def thread_function(dataset):
66
66
67
67
json_dataset = json .loads (uploaded_file .read ())
68
68
69
+ dataset_from_db = get_dataset (dataset_id = json_dataset ['dataset_id' ])
70
+ if dataset_from_db is not None and dataset_from_db ['clusters_computed' ] != len (dataset_from_db ['questions' ]):
71
+ return f'rejected file: { uploaded_file .name } , dataset still uploading'
72
+
69
73
Thread (target = thread_function , args = (json_dataset ,)).start ()
70
74
71
75
return f'uploaded file: { uploaded_file .name } successfully'
You can’t perform that action at this time.
0 commit comments