Skip to content

Commit 040a682

Browse files
committed
venv instruction fixed
1 parent dbbc0c9 commit 040a682

File tree

4 files changed

+3
-16
lines changed

4 files changed

+3
-16
lines changed

Application Project 1.pdf

134 KB
Binary file not shown.

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ After cloning the repository, you need to initiate your `virtualenv` and load th
2121

2222
```bash
2323
python -m venv toolkit_venv
24-
source toolkit_venv
24+
source toolkit_venv/bin/activate
2525
pip install -r requirements.txt
2626
```
2727

example/demo.ipynb

Lines changed: 2 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -9,22 +9,9 @@
99
},
1010
{
1111
"cell_type": "code",
12-
"execution_count": 1,
12+
"execution_count": null,
1313
"metadata": {},
14-
"outputs": [
15-
{
16-
"ename": "ModuleNotFoundError",
17-
"evalue": "No module named 'Turkish_Preprocessing'",
18-
"output_type": "error",
19-
"traceback": [
20-
"\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
21-
"\u001b[0;31mModuleNotFoundError\u001b[0m Traceback (most recent call last)",
22-
"\u001b[0;32m<ipython-input-1-abe21d06f4b3>\u001b[0m in \u001b[0;36m<module>\u001b[0;34m\u001b[0m\n\u001b[0;32m----> 1\u001b[0;31m \u001b[0;32mfrom\u001b[0m \u001b[0mmodules\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mtokenizer\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mTokenizer\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m 2\u001b[0m \u001b[0;32mfrom\u001b[0m \u001b[0mmodules\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0msplitter\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mSentenceSplitter\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 3\u001b[0m \u001b[0;32mfrom\u001b[0m \u001b[0mmodules\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mnormalization\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mNormalization\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 4\u001b[0m \u001b[0;32mfrom\u001b[0m \u001b[0mmodules\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mstopwords\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mStopwordRemoval\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 5\u001b[0m \u001b[0;32mfrom\u001b[0m \u001b[0mmodules\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mstemmer\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mStemmer\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
23-
"\u001b[0;32m~/Documents/Courses/Fall 2021/CMPE561 - NATURAL LANGUAGE PROCESSING/Application Projects/Turkish_Preprocessing/modules/tokenizer.py\u001b[0m in \u001b[0;36m<module>\u001b[0;34m\u001b[0m\n\u001b[1;32m 4\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mre\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 5\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mwarnings\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m----> 6\u001b[0;31m \u001b[0;32mfrom\u001b[0m \u001b[0mTurkish_Preprocessing\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mutils_\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mimport_corpora\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mimport_mwe_list\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mimport_abbreviations\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m 7\u001b[0m \u001b[0;32mfrom\u001b[0m \u001b[0mTurkish_Preprocessing\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mmodules\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mnormalization\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mNormalization\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 8\u001b[0m \u001b[0;32mimport\u001b[0m \u001b[0mnumpy\u001b[0m \u001b[0;32mas\u001b[0m \u001b[0mnp\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
24-
"\u001b[0;31mModuleNotFoundError\u001b[0m: No module named 'Turkish_Preprocessing'"
25-
]
26-
}
27-
],
14+
"outputs": [],
2815
"source": [
2916
"from Turkish_Preprocessing.modules.tokenizer import Tokenizer\n",
3017
"from Turkish_Preprocessing.modules.splitter import SentenceSplitter\n",

rapor.pdf

114 KB
Binary file not shown.

0 commit comments

Comments
 (0)