I'm facing the same issue on Windows :/
I'll search to see if I can see where this come from
Ok so I just tried again on my computer.
The same issue. I really think that it's just a import performance issue.
I not allowed to send the import package as It contain many sensitive data.
Through I tried to delete all sensitive data and import the file, and it works !!! But yes, the import package is only 100ko now so...
Didn't found any workaround for now.. Still the same issue, but now I'm sure it's not related to NFS
Actually, when I don't persist data on NFS server (no volume on docker container) it take a minute to import data.
I benchmarked communications between container and NFS Server :
root@ac2bf523cd00:/app# time dd if=/dev/zero of=data/testfile bs=16k count=128k 131072+0 records in 131072+0 records out 2147483648 bytes (2.1 GB, 2.0 GiB) copied, 18.434 s, 116 MB/s real 0m19.292s user 0m0.029s sys 0m2.351s
I don't have exact data on editing big entities but yet it's pretty fast.
Also just tried to import 5.4Mo file and still the same issue
I'll give a shot at MongoDB
Looks like your connection to jsreport forum was lost, please wait while we try to reconnect.