-
-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
missing file in Data stated in constants.py #5
Comments
I believe the data files are generated from |
There are data files needed to train and test files in the current version of my python code. Some of these are data used to train/test models (my.txt, BEST data, etc), and some of these are data files generated by my code that are used at the evaluation time as well, such as By the end of my internship, I shared a google drive folder with Shane (called Dictionary Segmentation) that has all these files. I just shared the same folder with Frank. |
Should we at least check in github all the files needed to TEST / Eval the segmentation. I think we should not check in all the data which train the model, but for anything that are needed run AFTER the training, should we check them into github? |
We shouldn't check in data files that are strongly coupled with the training data. Instead, it would be better design if the code didn't need those files to exist at all. Ideally the code should be able to pull what it needs directly from the model files. |
I think the main data file that we need for the evaluation is the dictionary that has grapheme clusters in it (e.g. |
OK, so I think we should probably just check the ratio files into the repo then. Otherwise, someone who downloads the repo won't be able to run the code. Does that sound okay to you @SahandFarhoodi ? |
Yes, I think that's the best solution. |
I try to run the basic test but mostly failed
It seems you coded some data path in constatnts.py but those files do not exist.
ftang@ftang4:~/lstm_word_segmentation$ python3 test/test_helpers.py
Traceback (most recent call last):
File "test/test_helpers.py", line 3, in
from lstm_word_segmentation.helpers import is_ascii, diff_strings, sigmoid
File "/usr/local/google/home/ftang/lstm_word_segmentation/lstm_word_segmentation/helpers.py", line 2, in
from . import constants
File "/usr/local/google/home/ftang/lstm_word_segmentation/lstm_word_segmentation/constants.py", line 7, in
THAI_GRAPH_CLUST_RATIO = np.load(str(path), allow_pickle=True).item()
File "/usr/lib/python3/dist-packages/numpy/lib/npyio.py", line 416, in load
fid = stack.enter_context(open(os_fspath(file), "rb"))
FileNotFoundError: [Errno 2] No such file or directory: '/usr/local/google/home/ftang/lstm_word_segmentation/Data/Thai_graph_clust_ratio.npy'
ft
The text was updated successfully, but these errors were encountered: