From 77cb00c824c4c00bb06cdd618be14af5cc1723a5 Mon Sep 17 00:00:00 2001 From: rduo1009 Date: Mon, 20 Jan 2025 21:32:41 +1100 Subject: [PATCH] =?UTF-8?q?=F0=9F=94=A8=20refactor:=20removing=20unneeded?= =?UTF-8?q?=20type=20stubs?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- stubs/aenum/__init__.pyi | 97 - stubs/aenum/_common.pyi | 127 -- stubs/aenum/_constant.pyi | 37 - stubs/aenum/_enum.pyi | 293 +-- stubs/aenum/_py3.pyi | 6 - stubs/aenum/_tuple.pyi | 64 - stubs/aenum/allowlist | 193 +- stubs/colors/__init__.pyi | 1 - stubs/colors/allowlist | 28 + stubs/colors/colors.pyi | 32 +- stubs/colors/csscolors.pyi | 8 - stubs/dill/__init__.pyi | 39 - stubs/dill/_dill.pyi | 147 +- stubs/dill/_objects.pyi | 26 - stubs/dill/_shims.pyi | 28 - stubs/dill/allowlist | 31 +- stubs/dill/detect.pyi | 56 - stubs/dill/logger.pyi | 58 - stubs/dill/objtypes.pyi | 1 - stubs/dill/pointers.pyi | 24 - stubs/dill/session.pyi | 53 - stubs/dill/settings.pyi | 5 - stubs/dill/source.pyi | 71 - stubs/dill/temp.pyi | 26 - stubs/lemminflect/__init__.pyi | 26 +- stubs/lemminflect/allowlist | 45 +- stubs/lemminflect/codecs/FormsTable.pyi | 7 - .../lemminflect/codecs/InflTCorpFileCodec.pyi | 21 - .../lemminflect/codecs/InflectionLUCodec.pyi | 40 - stubs/lemminflect/codecs/LemmaLUCodec.pyi | 22 - .../codecs/LemmaTCorpFileCodec.pyi | 29 - stubs/lemminflect/codecs/OverridesCodec.pyi | 12 - stubs/lemminflect/codecs/__init__.pyi | 1 - stubs/lemminflect/config.pyi | 26 - stubs/lemminflect/core/InflectionRules.pyi | 26 - stubs/lemminflect/core/Inflections.pyi | 48 - stubs/lemminflect/core/Lemmatizer.pyi | 43 - stubs/lemminflect/core/LemmatizerRules.pyi | 12 - stubs/lemminflect/core/LexicalUtils.pyi | 19 - stubs/lemminflect/core/__init__.pyi | 1 - stubs/lemminflect/kmodels/KInfer.pyi | 38 - stubs/lemminflect/kmodels/KerasModel.pyi | 15 - stubs/lemminflect/kmodels/ModelInfl.pyi | 23 - stubs/lemminflect/kmodels/ModelInflInData.pyi | 14 - stubs/lemminflect/kmodels/ModelLemma.pyi | 24 - .../lemminflect/kmodels/ModelLemmaClasses.pyi | 21 - .../lemminflect/kmodels/ModelLemmaInData.pyi | 23 - stubs/lemminflect/kmodels/__init__.pyi | 1 - stubs/lemminflect/slexicon/SKey.pyi | 48 - .../lemminflect/slexicon/SPECIALISTEntry.pyi | 44 - .../slexicon/SPECIALISTExtractor.pyi | 11 - stubs/lemminflect/slexicon/__init__.pyi | 1 - stubs/lemminflect/utils/CorpusUtils.pyi | 7 - stubs/lemminflect/utils/DataContainer.pyi | 9 - stubs/lemminflect/utils/ProgressBar.pyi | 6 - stubs/lemminflect/utils/Singleton.pyi | 12 - stubs/lemminflect/utils/Unigrams.pyi | 25 - stubs/lemminflect/utils/__init__.pyi | 1 - stubs/nltk/__init__.pyi | 124 +- stubs/nltk/allowlist | 2017 +++-------------- stubs/nltk/app/__init__.pyi | 1 - stubs/nltk/app/chartparser_app.pyi | 169 -- stubs/nltk/app/chunkparser_app.pyi | 52 - stubs/nltk/app/collocations_app.pyi | 67 - stubs/nltk/app/concordance_app.pyi | 97 - stubs/nltk/app/nemo_app.pyi | 44 - stubs/nltk/app/rdparser_app.pyi | 38 - stubs/nltk/app/srparser_app.pyi | 34 - stubs/nltk/app/wordfreq_app.pyi | 3 - stubs/nltk/app/wordnet_app.pyi | 31 - stubs/nltk/book.pyi | 36 - stubs/nltk/ccg/__init__.pyi | 30 - stubs/nltk/ccg/api.pyi | 76 - stubs/nltk/ccg/chart.pyi | 117 - stubs/nltk/ccg/combinator.pyi | 111 - stubs/nltk/ccg/lexicon.pyi | 61 - stubs/nltk/ccg/logic.pyi | 14 - stubs/nltk/chat/__init__.pyi | 12 - stubs/nltk/chat/eliza.pyi | 10 - stubs/nltk/chat/iesha.pyi | 10 - stubs/nltk/chat/rude.pyi | 10 - stubs/nltk/chat/suntsu.pyi | 10 - stubs/nltk/chat/util.pyi | 10 - stubs/nltk/chat/zen.pyi | 10 - stubs/nltk/chunk/__init__.pyi | 22 - stubs/nltk/chunk/api.pyi | 10 - stubs/nltk/chunk/named_entity.pyi | 41 - stubs/nltk/chunk/regexp.pyi | 117 - stubs/nltk/chunk/util.pyi | 58 - stubs/nltk/classify/__init__.pyi | 60 - stubs/nltk/classify/api.pyi | 17 - stubs/nltk/classify/decisiontree.pyi | 76 - stubs/nltk/classify/maxent.pyi | 84 - stubs/nltk/classify/megam.pyi | 23 - stubs/nltk/classify/naivebayes.pyi | 33 - stubs/nltk/classify/positivenaivebayes.pyi | 25 - stubs/nltk/classify/rte_classify.pyi | 7 - stubs/nltk/classify/scikitlearn.pyi | 17 - stubs/nltk/classify/senna.pyi | 11 - stubs/nltk/classify/tadm.pyi | 18 - stubs/nltk/classify/textcat.pyi | 17 - stubs/nltk/classify/util.pyi | 35 - stubs/nltk/classify/weka.pyi | 38 - stubs/nltk/collocations.pyi | 97 - stubs/nltk/compat.pyi | 6 - stubs/nltk/corpus/__init__.pyi | 83 - stubs/nltk/corpus/europarl_raw.pyi | 14 - stubs/nltk/corpus/reader/__init__.pyi | 197 -- stubs/nltk/corpus/reader/aligned.pyi | 50 - stubs/nltk/corpus/reader/api.pyi | 74 +- stubs/nltk/corpus/reader/bcp47.pyi | 22 - stubs/nltk/corpus/reader/bnc.pyi | 68 - stubs/nltk/corpus/reader/bracket_parse.pyi | 48 - .../nltk/corpus/reader/categorized_sents.pyi | 26 - stubs/nltk/corpus/reader/chasen.pyi | 41 - stubs/nltk/corpus/reader/childes.pyi | 82 - stubs/nltk/corpus/reader/chunked.pyi | 73 - stubs/nltk/corpus/reader/cmudict.pyi | 12 - .../nltk/corpus/reader/comparative_sents.pyi | 45 - stubs/nltk/corpus/reader/conll.pyi | 121 - stubs/nltk/corpus/reader/crubadan.pyi | 18 - stubs/nltk/corpus/reader/dependency.pyi | 44 - stubs/nltk/corpus/reader/framenet.pyi | 141 -- stubs/nltk/corpus/reader/framenet_typed.pyi | 163 -- stubs/nltk/corpus/reader/ieer.pyi | 25 - stubs/nltk/corpus/reader/indian.pyi | 31 - stubs/nltk/corpus/reader/ipipan.pyi | 3 - stubs/nltk/corpus/reader/knbc.pyi | 31 - stubs/nltk/corpus/reader/lin.pyi | 21 - stubs/nltk/corpus/reader/markdown.pyi | 157 -- stubs/nltk/corpus/reader/mte.pyi | 90 - stubs/nltk/corpus/reader/mte_typed.pyi | 100 - stubs/nltk/corpus/reader/nkjp.pyi | 3 - stubs/nltk/corpus/reader/nombank.pyi | 21 - stubs/nltk/corpus/reader/nps_chat.pyi | 30 - stubs/nltk/corpus/reader/opinion_lexicon.pyi | 14 - stubs/nltk/corpus/reader/panlex_lite.pyi | 25 - stubs/nltk/corpus/reader/panlex_swadesh.pyi | 30 - stubs/nltk/corpus/reader/pl196x.pyi | 76 - stubs/nltk/corpus/reader/plaintext.pyi | 34 - stubs/nltk/corpus/reader/ppattach.pyi | 25 - stubs/nltk/corpus/reader/propbank.pyi | 115 - stubs/nltk/corpus/reader/pros_cons.pyi | 25 - stubs/nltk/corpus/reader/reviews.pyi | 46 - stubs/nltk/corpus/reader/rte.pyi | 31 - stubs/nltk/corpus/reader/semcor.pyi | 51 - stubs/nltk/corpus/reader/senseval.pyi | 25 - stubs/nltk/corpus/reader/sentiwordnet.pyi | 24 - stubs/nltk/corpus/reader/sinica_treebank.pyi | 13 - stubs/nltk/corpus/reader/string_category.pyi | 14 - stubs/nltk/corpus/reader/switchboard.pyi | 24 - stubs/nltk/corpus/reader/tagged.pyi | 92 - stubs/nltk/corpus/reader/timit.pyi | 86 - stubs/nltk/corpus/reader/toolbox.pyi | 25 - stubs/nltk/corpus/reader/twitter.pyi | 26 - stubs/nltk/corpus/reader/udhr.pyi | 11 - stubs/nltk/corpus/reader/util.pyi | 43 +- stubs/nltk/corpus/reader/verbnet.pyi | 37 - stubs/nltk/corpus/reader/wordlist.pyi | 36 - stubs/nltk/corpus/reader/wordnet.pyi | 32 +- stubs/nltk/corpus/reader/xmldocs.pyi | 33 - stubs/nltk/corpus/reader/ycoe.pyi | 38 - stubs/nltk/corpus/util.pyi | 9 +- stubs/nltk/data.pyi | 77 +- stubs/nltk/decorators.pyi | 6 - stubs/nltk/downloader.pyi | 225 -- stubs/nltk/draw/__init__.pyi | 26 - stubs/nltk/draw/cfg.pyi | 56 - stubs/nltk/draw/dispersion.pyi | 8 - stubs/nltk/draw/table.pyi | 192 -- stubs/nltk/draw/tree.pyi | 91 - stubs/nltk/draw/util.pyi | 213 -- stubs/nltk/featstruct.pyi | 277 --- stubs/nltk/grammar.pyi | 165 -- stubs/nltk/help.pyi | 7 - stubs/nltk/inference/__init__.pyi | 38 - stubs/nltk/inference/api.pyi | 143 -- stubs/nltk/inference/discourse.pyi | 93 - stubs/nltk/inference/mace.pyi | 42 - stubs/nltk/inference/nonmonotonic.pyi | 93 - stubs/nltk/inference/prover9.pyi | 87 - stubs/nltk/inference/resolution.pyi | 116 - stubs/nltk/inference/tableau.pyi | 129 -- stubs/nltk/internals.pyi | 187 -- stubs/nltk/jsontags.pyi | 22 - stubs/nltk/langnames.pyi | 19 - stubs/nltk/lazyimport.pyi | 11 - stubs/nltk/lm/__init__.pyi | 35 - stubs/nltk/lm/api.pyi | 53 - stubs/nltk/lm/counter.pyi | 16 - stubs/nltk/lm/models.pyi | 64 - stubs/nltk/lm/preprocessing.pyi | 14 - stubs/nltk/lm/smoothing.pyi | 42 - stubs/nltk/lm/util.pyi | 6 - stubs/nltk/lm/vocabulary.pyi | 24 - stubs/nltk/metrics/__init__.pyi | 82 - stubs/nltk/metrics/agreement.pyi | 57 - stubs/nltk/metrics/aline.pyi | 47 - stubs/nltk/metrics/association.pyi | 89 - stubs/nltk/metrics/confusionmatrix.pyi | 33 - stubs/nltk/metrics/distance.pyi | 29 - stubs/nltk/metrics/paice.pyi | 13 - stubs/nltk/metrics/scores.pyi | 20 - stubs/nltk/metrics/segmentation.pyi | 23 - stubs/nltk/metrics/spearman.pyi | 11 - stubs/nltk/misc/__init__.pyi | 4 - stubs/nltk/misc/babelfish.pyi | 1 - stubs/nltk/misc/chomsky.pyi | 6 - stubs/nltk/misc/minimalset.pyi | 13 - stubs/nltk/misc/sort.pyi | 7 - stubs/nltk/misc/wordfinder.pyi | 28 - stubs/nltk/parse/__init__.pyi | 127 -- stubs/nltk/parse/api.pyi | 18 - stubs/nltk/parse/bllip.pyi | 31 - stubs/nltk/parse/chart.pyi | 261 --- stubs/nltk/parse/corenlp.pyi | 103 - stubs/nltk/parse/dependencygraph.pyi | 64 - stubs/nltk/parse/earleychart.pyi | 176 -- stubs/nltk/parse/evaluate.pyi | 7 - stubs/nltk/parse/featurechart.pyi | 146 -- stubs/nltk/parse/generate.pyi | 14 - stubs/nltk/parse/malt.pyi | 57 - .../parse/nonprojectivedependencyparser.pyi | 64 - stubs/nltk/parse/pchart.pyi | 94 - .../nltk/parse/projectivedependencyparser.pyi | 57 - stubs/nltk/parse/recursivedescent.pyi | 32 - stubs/nltk/parse/shiftreduce.pyi | 31 - stubs/nltk/parse/stanford.pyi | 59 - stubs/nltk/parse/transitionparser.pyi | 49 - stubs/nltk/parse/util.pyi | 57 - stubs/nltk/parse/viterbi.pyi | 17 - stubs/nltk/probability.pyi | 302 --- stubs/nltk/sem/__init__.pyi | 74 - stubs/nltk/sem/boxer.pyi | 320 --- stubs/nltk/sem/chat80.pyi | 71 - stubs/nltk/sem/cooper_storage.pyi | 28 - stubs/nltk/sem/drt.pyi | 306 --- stubs/nltk/sem/drt_glue_demo.pyi | 31 - stubs/nltk/sem/evaluate.pyi | 119 - stubs/nltk/sem/glue.pyi | 146 -- stubs/nltk/sem/hole.pyi | 71 - stubs/nltk/sem/lfg.pyi | 16 - stubs/nltk/sem/linearlogic.pyi | 115 - stubs/nltk/sem/logic.pyi | 439 ---- stubs/nltk/sem/relextract.pyi | 31 - stubs/nltk/sem/skolemize.pyi | 45 - stubs/nltk/sem/util.pyi | 25 - stubs/nltk/sentiment/__init__.pyi | 6 - stubs/nltk/sentiment/sentiment_analyzer.pyi | 55 - stubs/nltk/sentiment/util.pyi | 73 - stubs/nltk/sentiment/vader.pyi | 58 - stubs/nltk/stem/__init__.pyi | 11 - stubs/nltk/stem/api.pyi | 7 - stubs/nltk/stem/arlstem.pyi | 39 - stubs/nltk/stem/arlstem2.pyi | 42 - stubs/nltk/stem/cistem.pyi | 19 - stubs/nltk/stem/isri.pyi | 32 - stubs/nltk/stem/lancaster.pyi | 14 - stubs/nltk/stem/porter.pyi | 19 - stubs/nltk/stem/regexp.pyi | 7 - stubs/nltk/stem/rslp.pyi | 12 - stubs/nltk/stem/snowball.pyi | 91 - stubs/nltk/stem/util.pyi | 8 - stubs/nltk/stem/wordnet.pyi | 10 - stubs/nltk/tabdata.pyi | 28 - stubs/nltk/tag/__init__.pyi | 92 - stubs/nltk/tag/api.pyi | 35 - stubs/nltk/tag/brill.pyi | 49 - stubs/nltk/tag/brill_trainer.pyi | 21 - stubs/nltk/tag/crf.pyi | 17 - stubs/nltk/tag/hmm.pyi | 138 -- stubs/nltk/tag/hunpos.pyi | 26 - stubs/nltk/tag/mapping.pyi | 8 - stubs/nltk/tag/perceptron.pyi | 55 - stubs/nltk/tag/senna.pyi | 20 - stubs/nltk/tag/sequential.pyi | 158 -- stubs/nltk/tag/stanford.pyi | 40 - stubs/nltk/tag/tnt.pyi | 28 - stubs/nltk/tag/util.pyi | 5 - stubs/nltk/tbl/__init__.pyi | 4 - stubs/nltk/tbl/api.pyi | 0 stubs/nltk/tbl/demo.pyi | 55 - stubs/nltk/tbl/erroranalysis.pyi | 5 - stubs/nltk/tbl/feature.pyi | 34 - stubs/nltk/tbl/rule.pyi | 39 - stubs/nltk/tbl/template.pyi | 35 - stubs/nltk/text.pyi | 92 - stubs/nltk/tgrep.pyi | 17 - stubs/nltk/tokenize/__init__.pyi | 87 - stubs/nltk/tokenize/api.pyi | 23 - stubs/nltk/tokenize/casual.pyi | 43 - stubs/nltk/tokenize/destructive.pyi | 28 - stubs/nltk/tokenize/legality_principle.pyi | 17 - stubs/nltk/tokenize/mwe.pyi | 11 - stubs/nltk/tokenize/nist.pyi | 40 - stubs/nltk/tokenize/punkt.pyi | 170 -- stubs/nltk/tokenize/regexp.pyi | 40 - stubs/nltk/tokenize/repp.pyi | 27 - stubs/nltk/tokenize/sexpr.pyi | 9 - stubs/nltk/tokenize/simple.pyi | 34 - stubs/nltk/tokenize/sonority_sequencing.pyi | 19 - stubs/nltk/tokenize/stanford.pyi | 25 - stubs/nltk/tokenize/stanford_segmenter.pyi | 41 - stubs/nltk/tokenize/texttiling.pyi | 48 - stubs/nltk/tokenize/toktok.pyi | 38 - stubs/nltk/tokenize/treebank.pyi | 38 - stubs/nltk/tokenize/util.pyi | 35 - stubs/nltk/toolbox.pyi | 63 - stubs/nltk/translate/__init__.pyi | 20 - stubs/nltk/translate/api.pyi | 49 - stubs/nltk/translate/bleu_score.pyi | 64 - stubs/nltk/translate/chrf_score.pyi | 27 - stubs/nltk/translate/gale_church.pyi | 44 - stubs/nltk/translate/gdfa.pyi | 8 - stubs/nltk/translate/gleu_score.pyi | 17 - stubs/nltk/translate/ibm1.pyi | 31 - stubs/nltk/translate/ibm2.pyi | 50 - stubs/nltk/translate/ibm3.pyi | 79 - stubs/nltk/translate/ibm4.pyi | 94 - stubs/nltk/translate/ibm5.pyi | 95 - stubs/nltk/translate/ibm_model.pyi | 91 - stubs/nltk/translate/meteor_score.pyi | 93 - stubs/nltk/translate/metrics.pyi | 7 - stubs/nltk/translate/nist_score.pyi | 15 - stubs/nltk/translate/phrase_based.pyi | 21 - stubs/nltk/translate/ribes_score.pyi | 40 - stubs/nltk/translate/stack_decoder.pyi | 61 - stubs/nltk/tree/__init__.pyi | 53 - stubs/nltk/tree/immutable.pyi | 47 - stubs/nltk/tree/parented.pyi | 51 - stubs/nltk/tree/parsing.pyi | 5 - stubs/nltk/tree/prettyprinter.pyi | 33 - stubs/nltk/tree/probabilistic.pyi | 19 - stubs/nltk/tree/transforms.pyi | 25 - stubs/nltk/tree/tree.pyi | 98 - stubs/nltk/treeprettyprinter.pyi | 6 - stubs/nltk/treetransforms.pyi | 7 - stubs/nltk/twitter/__init__.pyi | 22 - stubs/nltk/twitter/api.pyi | 39 - stubs/nltk/twitter/common.pyi | 34 - stubs/nltk/twitter/twitter_demo.pyi | 42 - stubs/nltk/twitter/twitterclient.pyi | 109 - stubs/nltk/twitter/util.pyi | 24 - stubs/nltk/util.pyi | 148 -- stubs/nltk/wsd.pyi | 11 - tox.ini | 2 +- 346 files changed, 622 insertions(+), 19074 deletions(-) delete mode 100644 stubs/aenum/_constant.pyi delete mode 100644 stubs/aenum/_py3.pyi delete mode 100644 stubs/aenum/_tuple.pyi create mode 100644 stubs/colors/allowlist delete mode 100644 stubs/colors/csscolors.pyi delete mode 100644 stubs/dill/_objects.pyi delete mode 100644 stubs/dill/_shims.pyi delete mode 100644 stubs/dill/detect.pyi delete mode 100644 stubs/dill/logger.pyi delete mode 100644 stubs/dill/objtypes.pyi delete mode 100644 stubs/dill/pointers.pyi delete mode 100644 stubs/dill/session.pyi delete mode 100644 stubs/dill/settings.pyi delete mode 100644 stubs/dill/source.pyi delete mode 100644 stubs/dill/temp.pyi delete mode 100644 stubs/lemminflect/codecs/FormsTable.pyi delete mode 100644 stubs/lemminflect/codecs/InflTCorpFileCodec.pyi delete mode 100644 stubs/lemminflect/codecs/InflectionLUCodec.pyi delete mode 100644 stubs/lemminflect/codecs/LemmaLUCodec.pyi delete mode 100644 stubs/lemminflect/codecs/LemmaTCorpFileCodec.pyi delete mode 100644 stubs/lemminflect/codecs/OverridesCodec.pyi delete mode 100644 stubs/lemminflect/codecs/__init__.pyi delete mode 100644 stubs/lemminflect/config.pyi delete mode 100644 stubs/lemminflect/core/InflectionRules.pyi delete mode 100644 stubs/lemminflect/core/Inflections.pyi delete mode 100644 stubs/lemminflect/core/Lemmatizer.pyi delete mode 100644 stubs/lemminflect/core/LemmatizerRules.pyi delete mode 100644 stubs/lemminflect/core/LexicalUtils.pyi delete mode 100644 stubs/lemminflect/core/__init__.pyi delete mode 100644 stubs/lemminflect/kmodels/KInfer.pyi delete mode 100644 stubs/lemminflect/kmodels/KerasModel.pyi delete mode 100644 stubs/lemminflect/kmodels/ModelInfl.pyi delete mode 100644 stubs/lemminflect/kmodels/ModelInflInData.pyi delete mode 100644 stubs/lemminflect/kmodels/ModelLemma.pyi delete mode 100644 stubs/lemminflect/kmodels/ModelLemmaClasses.pyi delete mode 100644 stubs/lemminflect/kmodels/ModelLemmaInData.pyi delete mode 100644 stubs/lemminflect/kmodels/__init__.pyi delete mode 100644 stubs/lemminflect/slexicon/SKey.pyi delete mode 100644 stubs/lemminflect/slexicon/SPECIALISTEntry.pyi delete mode 100644 stubs/lemminflect/slexicon/SPECIALISTExtractor.pyi delete mode 100644 stubs/lemminflect/slexicon/__init__.pyi delete mode 100644 stubs/lemminflect/utils/CorpusUtils.pyi delete mode 100644 stubs/lemminflect/utils/DataContainer.pyi delete mode 100644 stubs/lemminflect/utils/ProgressBar.pyi delete mode 100644 stubs/lemminflect/utils/Singleton.pyi delete mode 100644 stubs/lemminflect/utils/Unigrams.pyi delete mode 100644 stubs/lemminflect/utils/__init__.pyi delete mode 100644 stubs/nltk/app/__init__.pyi delete mode 100644 stubs/nltk/app/chartparser_app.pyi delete mode 100644 stubs/nltk/app/chunkparser_app.pyi delete mode 100644 stubs/nltk/app/collocations_app.pyi delete mode 100644 stubs/nltk/app/concordance_app.pyi delete mode 100644 stubs/nltk/app/nemo_app.pyi delete mode 100644 stubs/nltk/app/rdparser_app.pyi delete mode 100644 stubs/nltk/app/srparser_app.pyi delete mode 100644 stubs/nltk/app/wordfreq_app.pyi delete mode 100644 stubs/nltk/app/wordnet_app.pyi delete mode 100644 stubs/nltk/book.pyi delete mode 100644 stubs/nltk/ccg/__init__.pyi delete mode 100644 stubs/nltk/ccg/api.pyi delete mode 100644 stubs/nltk/ccg/chart.pyi delete mode 100644 stubs/nltk/ccg/combinator.pyi delete mode 100644 stubs/nltk/ccg/lexicon.pyi delete mode 100644 stubs/nltk/ccg/logic.pyi delete mode 100644 stubs/nltk/chat/__init__.pyi delete mode 100644 stubs/nltk/chat/eliza.pyi delete mode 100644 stubs/nltk/chat/iesha.pyi delete mode 100644 stubs/nltk/chat/rude.pyi delete mode 100644 stubs/nltk/chat/suntsu.pyi delete mode 100644 stubs/nltk/chat/util.pyi delete mode 100644 stubs/nltk/chat/zen.pyi delete mode 100644 stubs/nltk/chunk/__init__.pyi delete mode 100644 stubs/nltk/chunk/api.pyi delete mode 100644 stubs/nltk/chunk/named_entity.pyi delete mode 100644 stubs/nltk/chunk/regexp.pyi delete mode 100644 stubs/nltk/chunk/util.pyi delete mode 100644 stubs/nltk/classify/__init__.pyi delete mode 100644 stubs/nltk/classify/api.pyi delete mode 100644 stubs/nltk/classify/decisiontree.pyi delete mode 100644 stubs/nltk/classify/maxent.pyi delete mode 100644 stubs/nltk/classify/megam.pyi delete mode 100644 stubs/nltk/classify/naivebayes.pyi delete mode 100644 stubs/nltk/classify/positivenaivebayes.pyi delete mode 100644 stubs/nltk/classify/rte_classify.pyi delete mode 100644 stubs/nltk/classify/scikitlearn.pyi delete mode 100644 stubs/nltk/classify/senna.pyi delete mode 100644 stubs/nltk/classify/tadm.pyi delete mode 100644 stubs/nltk/classify/textcat.pyi delete mode 100644 stubs/nltk/classify/util.pyi delete mode 100644 stubs/nltk/classify/weka.pyi delete mode 100644 stubs/nltk/collocations.pyi delete mode 100644 stubs/nltk/compat.pyi delete mode 100644 stubs/nltk/corpus/europarl_raw.pyi delete mode 100644 stubs/nltk/corpus/reader/aligned.pyi delete mode 100644 stubs/nltk/corpus/reader/bcp47.pyi delete mode 100644 stubs/nltk/corpus/reader/bnc.pyi delete mode 100644 stubs/nltk/corpus/reader/bracket_parse.pyi delete mode 100644 stubs/nltk/corpus/reader/categorized_sents.pyi delete mode 100644 stubs/nltk/corpus/reader/chasen.pyi delete mode 100644 stubs/nltk/corpus/reader/childes.pyi delete mode 100644 stubs/nltk/corpus/reader/chunked.pyi delete mode 100644 stubs/nltk/corpus/reader/cmudict.pyi delete mode 100644 stubs/nltk/corpus/reader/comparative_sents.pyi delete mode 100644 stubs/nltk/corpus/reader/conll.pyi delete mode 100644 stubs/nltk/corpus/reader/crubadan.pyi delete mode 100644 stubs/nltk/corpus/reader/dependency.pyi delete mode 100644 stubs/nltk/corpus/reader/framenet.pyi delete mode 100644 stubs/nltk/corpus/reader/framenet_typed.pyi delete mode 100644 stubs/nltk/corpus/reader/ieer.pyi delete mode 100644 stubs/nltk/corpus/reader/indian.pyi delete mode 100644 stubs/nltk/corpus/reader/ipipan.pyi delete mode 100644 stubs/nltk/corpus/reader/knbc.pyi delete mode 100644 stubs/nltk/corpus/reader/lin.pyi delete mode 100644 stubs/nltk/corpus/reader/markdown.pyi delete mode 100644 stubs/nltk/corpus/reader/mte.pyi delete mode 100644 stubs/nltk/corpus/reader/mte_typed.pyi delete mode 100644 stubs/nltk/corpus/reader/nkjp.pyi delete mode 100644 stubs/nltk/corpus/reader/nombank.pyi delete mode 100644 stubs/nltk/corpus/reader/nps_chat.pyi delete mode 100644 stubs/nltk/corpus/reader/opinion_lexicon.pyi delete mode 100644 stubs/nltk/corpus/reader/panlex_lite.pyi delete mode 100644 stubs/nltk/corpus/reader/panlex_swadesh.pyi delete mode 100644 stubs/nltk/corpus/reader/pl196x.pyi delete mode 100644 stubs/nltk/corpus/reader/plaintext.pyi delete mode 100644 stubs/nltk/corpus/reader/ppattach.pyi delete mode 100644 stubs/nltk/corpus/reader/propbank.pyi delete mode 100644 stubs/nltk/corpus/reader/pros_cons.pyi delete mode 100644 stubs/nltk/corpus/reader/reviews.pyi delete mode 100644 stubs/nltk/corpus/reader/rte.pyi delete mode 100644 stubs/nltk/corpus/reader/semcor.pyi delete mode 100644 stubs/nltk/corpus/reader/senseval.pyi delete mode 100644 stubs/nltk/corpus/reader/sentiwordnet.pyi delete mode 100644 stubs/nltk/corpus/reader/sinica_treebank.pyi delete mode 100644 stubs/nltk/corpus/reader/string_category.pyi delete mode 100644 stubs/nltk/corpus/reader/switchboard.pyi delete mode 100644 stubs/nltk/corpus/reader/tagged.pyi delete mode 100644 stubs/nltk/corpus/reader/timit.pyi delete mode 100644 stubs/nltk/corpus/reader/toolbox.pyi delete mode 100644 stubs/nltk/corpus/reader/twitter.pyi delete mode 100644 stubs/nltk/corpus/reader/udhr.pyi delete mode 100644 stubs/nltk/corpus/reader/verbnet.pyi delete mode 100644 stubs/nltk/corpus/reader/wordlist.pyi delete mode 100644 stubs/nltk/corpus/reader/xmldocs.pyi delete mode 100644 stubs/nltk/corpus/reader/ycoe.pyi delete mode 100644 stubs/nltk/decorators.pyi delete mode 100644 stubs/nltk/downloader.pyi delete mode 100644 stubs/nltk/draw/__init__.pyi delete mode 100644 stubs/nltk/draw/cfg.pyi delete mode 100644 stubs/nltk/draw/dispersion.pyi delete mode 100644 stubs/nltk/draw/table.pyi delete mode 100644 stubs/nltk/draw/tree.pyi delete mode 100644 stubs/nltk/draw/util.pyi delete mode 100644 stubs/nltk/featstruct.pyi delete mode 100644 stubs/nltk/grammar.pyi delete mode 100644 stubs/nltk/help.pyi delete mode 100644 stubs/nltk/inference/__init__.pyi delete mode 100644 stubs/nltk/inference/api.pyi delete mode 100644 stubs/nltk/inference/discourse.pyi delete mode 100644 stubs/nltk/inference/mace.pyi delete mode 100644 stubs/nltk/inference/nonmonotonic.pyi delete mode 100644 stubs/nltk/inference/prover9.pyi delete mode 100644 stubs/nltk/inference/resolution.pyi delete mode 100644 stubs/nltk/inference/tableau.pyi delete mode 100644 stubs/nltk/internals.pyi delete mode 100644 stubs/nltk/jsontags.pyi delete mode 100644 stubs/nltk/langnames.pyi delete mode 100644 stubs/nltk/lazyimport.pyi delete mode 100644 stubs/nltk/lm/__init__.pyi delete mode 100644 stubs/nltk/lm/api.pyi delete mode 100644 stubs/nltk/lm/counter.pyi delete mode 100644 stubs/nltk/lm/models.pyi delete mode 100644 stubs/nltk/lm/preprocessing.pyi delete mode 100644 stubs/nltk/lm/smoothing.pyi delete mode 100644 stubs/nltk/lm/util.pyi delete mode 100644 stubs/nltk/lm/vocabulary.pyi delete mode 100644 stubs/nltk/metrics/__init__.pyi delete mode 100644 stubs/nltk/metrics/agreement.pyi delete mode 100644 stubs/nltk/metrics/aline.pyi delete mode 100644 stubs/nltk/metrics/association.pyi delete mode 100644 stubs/nltk/metrics/confusionmatrix.pyi delete mode 100644 stubs/nltk/metrics/distance.pyi delete mode 100644 stubs/nltk/metrics/paice.pyi delete mode 100644 stubs/nltk/metrics/scores.pyi delete mode 100644 stubs/nltk/metrics/segmentation.pyi delete mode 100644 stubs/nltk/metrics/spearman.pyi delete mode 100644 stubs/nltk/misc/__init__.pyi delete mode 100644 stubs/nltk/misc/babelfish.pyi delete mode 100644 stubs/nltk/misc/chomsky.pyi delete mode 100644 stubs/nltk/misc/minimalset.pyi delete mode 100644 stubs/nltk/misc/sort.pyi delete mode 100644 stubs/nltk/misc/wordfinder.pyi delete mode 100644 stubs/nltk/parse/__init__.pyi delete mode 100644 stubs/nltk/parse/api.pyi delete mode 100644 stubs/nltk/parse/bllip.pyi delete mode 100644 stubs/nltk/parse/chart.pyi delete mode 100644 stubs/nltk/parse/corenlp.pyi delete mode 100644 stubs/nltk/parse/dependencygraph.pyi delete mode 100644 stubs/nltk/parse/earleychart.pyi delete mode 100644 stubs/nltk/parse/evaluate.pyi delete mode 100644 stubs/nltk/parse/featurechart.pyi delete mode 100644 stubs/nltk/parse/generate.pyi delete mode 100644 stubs/nltk/parse/malt.pyi delete mode 100644 stubs/nltk/parse/nonprojectivedependencyparser.pyi delete mode 100644 stubs/nltk/parse/pchart.pyi delete mode 100644 stubs/nltk/parse/projectivedependencyparser.pyi delete mode 100644 stubs/nltk/parse/recursivedescent.pyi delete mode 100644 stubs/nltk/parse/shiftreduce.pyi delete mode 100644 stubs/nltk/parse/stanford.pyi delete mode 100644 stubs/nltk/parse/transitionparser.pyi delete mode 100644 stubs/nltk/parse/util.pyi delete mode 100644 stubs/nltk/parse/viterbi.pyi delete mode 100644 stubs/nltk/probability.pyi delete mode 100644 stubs/nltk/sem/__init__.pyi delete mode 100644 stubs/nltk/sem/boxer.pyi delete mode 100644 stubs/nltk/sem/chat80.pyi delete mode 100644 stubs/nltk/sem/cooper_storage.pyi delete mode 100644 stubs/nltk/sem/drt.pyi delete mode 100644 stubs/nltk/sem/drt_glue_demo.pyi delete mode 100644 stubs/nltk/sem/evaluate.pyi delete mode 100644 stubs/nltk/sem/glue.pyi delete mode 100644 stubs/nltk/sem/hole.pyi delete mode 100644 stubs/nltk/sem/lfg.pyi delete mode 100644 stubs/nltk/sem/linearlogic.pyi delete mode 100644 stubs/nltk/sem/logic.pyi delete mode 100644 stubs/nltk/sem/relextract.pyi delete mode 100644 stubs/nltk/sem/skolemize.pyi delete mode 100644 stubs/nltk/sem/util.pyi delete mode 100644 stubs/nltk/sentiment/__init__.pyi delete mode 100644 stubs/nltk/sentiment/sentiment_analyzer.pyi delete mode 100644 stubs/nltk/sentiment/util.pyi delete mode 100644 stubs/nltk/sentiment/vader.pyi delete mode 100644 stubs/nltk/stem/__init__.pyi delete mode 100644 stubs/nltk/stem/api.pyi delete mode 100644 stubs/nltk/stem/arlstem.pyi delete mode 100644 stubs/nltk/stem/arlstem2.pyi delete mode 100644 stubs/nltk/stem/cistem.pyi delete mode 100644 stubs/nltk/stem/isri.pyi delete mode 100644 stubs/nltk/stem/lancaster.pyi delete mode 100644 stubs/nltk/stem/porter.pyi delete mode 100644 stubs/nltk/stem/regexp.pyi delete mode 100644 stubs/nltk/stem/rslp.pyi delete mode 100644 stubs/nltk/stem/snowball.pyi delete mode 100644 stubs/nltk/stem/util.pyi delete mode 100644 stubs/nltk/stem/wordnet.pyi delete mode 100644 stubs/nltk/tabdata.pyi delete mode 100644 stubs/nltk/tag/__init__.pyi delete mode 100644 stubs/nltk/tag/api.pyi delete mode 100644 stubs/nltk/tag/brill.pyi delete mode 100644 stubs/nltk/tag/brill_trainer.pyi delete mode 100644 stubs/nltk/tag/crf.pyi delete mode 100644 stubs/nltk/tag/hmm.pyi delete mode 100644 stubs/nltk/tag/hunpos.pyi delete mode 100644 stubs/nltk/tag/mapping.pyi delete mode 100644 stubs/nltk/tag/perceptron.pyi delete mode 100644 stubs/nltk/tag/senna.pyi delete mode 100644 stubs/nltk/tag/sequential.pyi delete mode 100644 stubs/nltk/tag/stanford.pyi delete mode 100644 stubs/nltk/tag/tnt.pyi delete mode 100644 stubs/nltk/tag/util.pyi delete mode 100644 stubs/nltk/tbl/__init__.pyi delete mode 100644 stubs/nltk/tbl/api.pyi delete mode 100644 stubs/nltk/tbl/demo.pyi delete mode 100644 stubs/nltk/tbl/erroranalysis.pyi delete mode 100644 stubs/nltk/tbl/feature.pyi delete mode 100644 stubs/nltk/tbl/rule.pyi delete mode 100644 stubs/nltk/tbl/template.pyi delete mode 100644 stubs/nltk/text.pyi delete mode 100644 stubs/nltk/tgrep.pyi delete mode 100644 stubs/nltk/tokenize/__init__.pyi delete mode 100644 stubs/nltk/tokenize/api.pyi delete mode 100644 stubs/nltk/tokenize/casual.pyi delete mode 100644 stubs/nltk/tokenize/destructive.pyi delete mode 100644 stubs/nltk/tokenize/legality_principle.pyi delete mode 100644 stubs/nltk/tokenize/mwe.pyi delete mode 100644 stubs/nltk/tokenize/nist.pyi delete mode 100644 stubs/nltk/tokenize/punkt.pyi delete mode 100644 stubs/nltk/tokenize/regexp.pyi delete mode 100644 stubs/nltk/tokenize/repp.pyi delete mode 100644 stubs/nltk/tokenize/sexpr.pyi delete mode 100644 stubs/nltk/tokenize/simple.pyi delete mode 100644 stubs/nltk/tokenize/sonority_sequencing.pyi delete mode 100644 stubs/nltk/tokenize/stanford.pyi delete mode 100644 stubs/nltk/tokenize/stanford_segmenter.pyi delete mode 100644 stubs/nltk/tokenize/texttiling.pyi delete mode 100644 stubs/nltk/tokenize/toktok.pyi delete mode 100644 stubs/nltk/tokenize/treebank.pyi delete mode 100644 stubs/nltk/tokenize/util.pyi delete mode 100644 stubs/nltk/toolbox.pyi delete mode 100644 stubs/nltk/translate/__init__.pyi delete mode 100644 stubs/nltk/translate/api.pyi delete mode 100644 stubs/nltk/translate/bleu_score.pyi delete mode 100644 stubs/nltk/translate/chrf_score.pyi delete mode 100644 stubs/nltk/translate/gale_church.pyi delete mode 100644 stubs/nltk/translate/gdfa.pyi delete mode 100644 stubs/nltk/translate/gleu_score.pyi delete mode 100644 stubs/nltk/translate/ibm1.pyi delete mode 100644 stubs/nltk/translate/ibm2.pyi delete mode 100644 stubs/nltk/translate/ibm3.pyi delete mode 100644 stubs/nltk/translate/ibm4.pyi delete mode 100644 stubs/nltk/translate/ibm5.pyi delete mode 100644 stubs/nltk/translate/ibm_model.pyi delete mode 100644 stubs/nltk/translate/meteor_score.pyi delete mode 100644 stubs/nltk/translate/metrics.pyi delete mode 100644 stubs/nltk/translate/nist_score.pyi delete mode 100644 stubs/nltk/translate/phrase_based.pyi delete mode 100644 stubs/nltk/translate/ribes_score.pyi delete mode 100644 stubs/nltk/translate/stack_decoder.pyi delete mode 100644 stubs/nltk/tree/__init__.pyi delete mode 100644 stubs/nltk/tree/immutable.pyi delete mode 100644 stubs/nltk/tree/parented.pyi delete mode 100644 stubs/nltk/tree/parsing.pyi delete mode 100644 stubs/nltk/tree/prettyprinter.pyi delete mode 100644 stubs/nltk/tree/probabilistic.pyi delete mode 100644 stubs/nltk/tree/transforms.pyi delete mode 100644 stubs/nltk/tree/tree.pyi delete mode 100644 stubs/nltk/treeprettyprinter.pyi delete mode 100644 stubs/nltk/treetransforms.pyi delete mode 100644 stubs/nltk/twitter/__init__.pyi delete mode 100644 stubs/nltk/twitter/api.pyi delete mode 100644 stubs/nltk/twitter/common.pyi delete mode 100644 stubs/nltk/twitter/twitter_demo.pyi delete mode 100644 stubs/nltk/twitter/twitterclient.pyi delete mode 100644 stubs/nltk/twitter/util.pyi delete mode 100644 stubs/nltk/util.pyi delete mode 100644 stubs/nltk/wsd.pyi diff --git a/stubs/aenum/__init__.pyi b/stubs/aenum/__init__.pyi index c88978e..9711ea7 100644 --- a/stubs/aenum/__init__.pyi +++ b/stubs/aenum/__init__.pyi @@ -1,99 +1,2 @@ -from typing import NamedTuple as NamedTuple - from ._common import * -from ._constant import * from ._enum import * -from ._tuple import * - -__all__ = [ - "NamedConstant", - "Constant", - "constant", - "skip", - "nonmember", - "member", - "no_arg", - "Member", - "NonMember", - "bin", - "Enum", - "IntEnum", - "AutoNumberEnum", - "OrderedEnum", - "UniqueEnum", - "StrEnum", - "UpperStrEnum", - "LowerStrEnum", - "ReprEnum", - "Flag", - "IntFlag", - "enum_property", - "AddValue", - "MagicValue", - "MultiValue", - "NoAlias", - "Unique", - "AddValueEnum", - "MultiValueEnum", - "NoAliasEnum", - "enum", - "extend_enum", - "unique", - "property", - "NamedTuple", - "_reduce_ex_by_name", - "FlagBoundary", - "STRICT", - "CONFORM", - "EJECT", - "KEEP", - "add_stdlib_integration", - "remove_stdlib_integration", - "AutoEnum", -] - -# Names in __all__ with no definition: -# AddValue -# AddValueEnum -# AutoEnum -# AutoNumberEnum -# CONFORM -# Constant -# EJECT -# Enum -# Flag -# FlagBoundary -# IntEnum -# IntFlag -# KEEP -# LowerStrEnum -# MagicValue -# Member -# MultiValue -# MultiValueEnum -# NamedConstant -# NamedTuple -# NoAlias -# NoAliasEnum -# NonMember -# OrderedEnum -# ReprEnum -# STRICT -# StrEnum -# Unique -# UniqueEnum -# UpperStrEnum -# _reduce_ex_by_name -# add_stdlib_integration -# bin -# constant -# enum -# enum_property -# extend_enum -# member -# no_arg -# nonmember -# property -# remove_stdlib_integration -# skip -# unique diff --git a/stubs/aenum/_common.pyi b/stubs/aenum/_common.pyi index 7745bd3..c7a71b8 100644 --- a/stubs/aenum/_common.pyi +++ b/stubs/aenum/_common.pyi @@ -1,116 +1,7 @@ from collections import OrderedDict as OrderedDict -from operator import abs as _abs_ -from operator import add as _add_ -from operator import and_ as _and_ -from operator import floordiv as _floordiv_ -from operator import inv as _inv_ -from operator import lshift as _lshift_ -from operator import mod as _mod_ -from operator import mul as _mul_ -from operator import neg as _neg_ -from operator import or_ as _or_ -from operator import pos as _pos_ -from operator import pow as _pow_ -from operator import rshift as _rshift_ -from operator import sub as _sub_ -from operator import truediv as _truediv_ -from operator import xor as _xor_ from _typeshed import Incomplete -from ._py3 import * - -__all__ = [ - "pyver", - "PY2", - "PY2_6", - "PY3", - "PY3_3", - "PY3_4", - "PY3_5", - "PY3_6", - "PY3_7", - "PY3_11", - "_or_", - "_and_", - "_xor_", - "_inv_", - "_abs_", - "_add_", - "_floordiv_", - "_lshift_", - "_rshift_", - "_mod_", - "_mul_", - "_neg_", - "_pos_", - "_pow_", - "_truediv_", - "_sub_", - "unicode", - "basestring", - "baseinteger", - "long", - "NoneType", - "_Addendum", - "is_descriptor", - "is_dunder", - "is_sunder", - "is_internal_class", - "is_private_name", - "get_attr_from_chain", - "_value", - "constant", - "make_class_unpicklable", - "bltin_property", - "skip", - "nonmember", - "member", - "Member", - "NonMember", - "OrderedDict", -] - -pyver: tuple[int, int] -PY2: bool -PY3: bool -PY2_6: tuple[int, int] -PY3_3: tuple[int, int] -PY3_4: tuple[int, int] -PY3_5: tuple[int, int] -PY3_6: tuple[int, int] -PY3_7: tuple[int, int] -PY3_11: tuple[int, int] -bltin_property = property -unicode = unicode -unicode = str -basestring: Incomplete -baseinteger: Incomplete -long = long -long = int -baseint = baseinteger -NoneType: Incomplete - -class _Addendum: - dict: Incomplete - ns: Incomplete - added: Incomplete - def __init__( - self, dict: Incomplete, doc: Incomplete, ns: Incomplete - ) -> None: ... - def __call__(self, func: Incomplete) -> Incomplete: ... - def __getitem__(self, name: Incomplete) -> Incomplete: ... - def __setitem__(self, name: Incomplete, value: Incomplete) -> None: ... - def resolve(self) -> Incomplete: ... - -def is_descriptor(obj: Incomplete) -> Incomplete: ... -def is_dunder(name: Incomplete) -> Incomplete: ... -def is_sunder(name: Incomplete) -> Incomplete: ... -def is_internal_class(cls_name: Incomplete, obj: Incomplete) -> Incomplete: ... -def is_private_name(cls_name: Incomplete, name: Incomplete) -> Incomplete: ... -def get_attr_from_chain(cls: Incomplete, attr: Incomplete) -> Incomplete: ... -def _value(obj: Incomplete) -> Incomplete: ... - class constant: value: Incomplete __doc__: Incomplete @@ -152,21 +43,3 @@ class constant: def __set_name__( self, ownerclass: Incomplete, name: Incomplete ) -> None: ... - -def make_class_unpicklable(obj: Incomplete) -> None: ... - -class NonMember: - value: Incomplete - def __init__(self, value: Incomplete) -> None: ... - def __get__( - self, instance: Incomplete, ownerclass: Incomplete | None = None - ) -> Incomplete: ... - -skip = NonMember -nonmember = NonMember - -class Member: - value: Incomplete - def __init__(self, value: Incomplete) -> None: ... - -member = Member diff --git a/stubs/aenum/_constant.pyi b/stubs/aenum/_constant.pyi deleted file mode 100644 index d015b82..0000000 --- a/stubs/aenum/_constant.pyi +++ /dev/null @@ -1,37 +0,0 @@ -from _typeshed import Incomplete - -from ._common import * - -__all__ = ["NamedConstant", "Constant"] - -NamedConstant: Incomplete - -class NamedConstantDict(dict): - def __init__(self) -> None: ... - def __setitem__(self, key: Incomplete, value: Incomplete) -> None: ... - -class NamedConstantMeta(type): - @classmethod - def __prepare__( - metacls: Incomplete, - cls: Incomplete, - bases: Incomplete, - **kwds: Incomplete, - ) -> Incomplete: ... - def __new__( - metacls: Incomplete, - cls: Incomplete, - bases: Incomplete, - clsdict: Incomplete, - ) -> Incomplete: ... - def __bool__(cls: Incomplete) -> bool: ... - def __delattr__(cls: Incomplete, attr: Incomplete) -> None: ... - def __iter__(cls: Incomplete) -> Incomplete: ... - def __reversed__(cls: Incomplete) -> Incomplete: ... - def __len__(cls: Incomplete) -> int: ... - __nonzero__ = __bool__ - def __setattr__( - cls: Incomplete, name: Incomplete, value: Incomplete - ) -> None: ... - -Constant = NamedConstant diff --git a/stubs/aenum/_enum.pyi b/stubs/aenum/_enum.pyi index 84f3051..aceb567 100644 --- a/stubs/aenum/_enum.pyi +++ b/stubs/aenum/_enum.pyi @@ -1,227 +1,9 @@ import enum as enum import sqlite3 as sqlite3 -from enum import EnumMeta as StdlibEnumMeta - -try: - from enum import FlagBoundary as FlagBoundary -except ImportError: - class FlagBoundary(StrEnum): - STRICT: Incomplete - CONFORM: Incomplete - EJECT: Incomplete - KEEP: Incomplete - -from types import DynamicClassAttribute from _typeshed import Incomplete from ._common import * -from ._constant import NamedConstant -from ._tuple import NamedTuple - -__all__ = [ - "bit_count", - "is_single_bit", - "bin", - "property", - "bits", - "AddValue", - "MagicValue", - "MultiValue", - "NoAlias", - "Unique", - "enum", - "auto", - "AddValueEnum", - "MultiValueEnum", - "NoAliasEnum", - "UniqueEnum", - "AutoNumberEnum", - "OrderedEnum", - "unique", - "no_arg", - "extend_enum", - "enum_property", - "EnumType", - "EnumMeta", - "EnumDict", - "Enum", - "IntEnum", - "StrEnum", - "Flag", - "IntFlag", - "LowerStrEnum", - "UpperStrEnum", - "ReprEnum", - "sqlite3", - "FlagBoundary", - "STRICT", - "CONFORM", - "EJECT", - "KEEP", - "add_stdlib_integration", - "remove_stdlib_integration", - "export", - "cls2module", - "_reduce_ex_by_name", - "show_flag_values", - "AutoEnum", -] -from typing import TYPE_CHECKING - -if TYPE_CHECKING: - from enum import Enum as Enum -else: - Enum: Incomplete - -RecursionError = RuntimeError -MagicValue: Incomplete -AddValue: Incomplete -MultiValue: Incomplete -NoAlias: Incomplete -Unique: Incomplete - -def export( - collection: Incomplete, namespace: Incomplete | None = None -) -> Incomplete: ... -def bit_count(num: Incomplete) -> Incomplete: ... -def is_single_bit(value: Incomplete) -> Incomplete: ... -def bin( - value: Incomplete, max_bits: Incomplete | None = None -) -> Incomplete: ... -def show_flag_values(value: Incomplete) -> Incomplete: ... - -base = DynamicClassAttribute -base = object # type: ignore[assignment] - -class property(base): - fget: Incomplete - fset: Incomplete - fdel: Incomplete - __doc__: Incomplete - overwrite_doc: Incomplete - __isabstractmethod__: Incomplete - def __init__( - self, - fget: Incomplete | None = None, - fset: Incomplete | None = None, - fdel: Incomplete | None = None, - doc: Incomplete | None = None, - ) -> None: ... - def getter(self, fget: Incomplete) -> Incomplete: ... - def setter(self, fset: Incomplete) -> Incomplete: ... - def deleter(self, fdel: Incomplete) -> Incomplete: ... - def __get__( - self, instance: Incomplete, ownerclass: Incomplete | None = None - ) -> Incomplete: ... - def __set__( - self, instance: Incomplete, value: Incomplete - ) -> Incomplete: ... - def __delete__(self, instance: Incomplete) -> Incomplete: ... - name: Incomplete - clsname: Incomplete - ownerclass: Incomplete - def __set_name__( - self, ownerclass: Incomplete, name: Incomplete - ) -> None: ... - -DynamicClassAttribute = property -enum_property = property - -class SentinelType(type): ... - -def bits(num: Incomplete) -> Incomplete: ... - -class EnumConstants(NamedConstant): - AddValue: Incomplete - MagicValue: Incomplete - MultiValue: Incomplete - NoAlias: Incomplete - Unique: Incomplete - -class ReprEnum: ... - -IntEnum: Incomplete - -class Flag: ... - -EJECT: Incomplete - -KEEP: Incomplete - -class enum: - name: Incomplete - def __init__(self, *args: Incomplete, **kwds: Incomplete) -> None: ... - def args(self) -> Incomplete: ... - def kwds(self) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - -class auto(enum): # type: ignore[valid-type] - enum_member: Incomplete - def __and__(self, other: Incomplete) -> Incomplete: ... - def __rand__(self, other: Incomplete) -> Incomplete: ... - def __invert__(self) -> Incomplete: ... - def __or__(self, other: Incomplete) -> Incomplete: ... - def __ror__(self, other: Incomplete) -> Incomplete: ... - def __xor__(self, other: Incomplete) -> Incomplete: ... - def __rxor__(self, other: Incomplete) -> Incomplete: ... - def __abs__(self) -> Incomplete: ... - def __add__(self, other: Incomplete) -> Incomplete: ... - def __radd__(self, other: Incomplete) -> Incomplete: ... - def __neg__(self) -> Incomplete: ... - def __pos__(self) -> Incomplete: ... - def __rdiv__(self, other: Incomplete) -> Incomplete: ... - def __floordiv__(self, other: Incomplete) -> Incomplete: ... - def __rfloordiv__(self, other: Incomplete) -> Incomplete: ... - def __truediv__(self, other: Incomplete) -> Incomplete: ... - def __rtruediv__(self, other: Incomplete) -> Incomplete: ... - def __lshift__(self, other: Incomplete) -> Incomplete: ... - def __rlshift__(self, other: Incomplete) -> Incomplete: ... - def __rshift__(self, other: Incomplete) -> Incomplete: ... - def __rrshift__(self, other: Incomplete) -> Incomplete: ... - def __mod__(self, other: Incomplete) -> Incomplete: ... - def __rmod__(self, other: Incomplete) -> Incomplete: ... - def __mul__(self, other: Incomplete) -> Incomplete: ... - def __rmul__(self, other: Incomplete) -> Incomplete: ... - def __pow__(self, other: Incomplete) -> Incomplete: ... - def __rpow__(self, other: Incomplete) -> Incomplete: ... - def __sub__(self, other: Incomplete) -> Incomplete: ... - def __rsub__(self, other: Incomplete) -> Incomplete: ... - def value(self) -> Incomplete: ... - @value.setter # type: ignore[attr-defined] - def value(self, value: Incomplete) -> None: ... - -class _EnumArgSpec(NamedTuple): - args: Incomplete - varargs: Incomplete - keywords: Incomplete - defaults: Incomplete - required: Incomplete - def __new__(cls: Incomplete, _new_func: Incomplete) -> Incomplete: ... - -class _proto_member: - value: Incomplete - def __init__(self, value: Incomplete) -> None: ... - def __set_name__( - self, enum_class: Incomplete, member_name: Incomplete - ) -> None: ... - -class EnumDict(dict): - def __init__( - self, - cls_name: Incomplete, - settings: Incomplete, - start: Incomplete, - constructor_init: Incomplete, - constructor_start: Incomplete, - constructor_boundary: Incomplete, - ) -> None: ... - def __getitem__(self, key: Incomplete) -> Incomplete: ... - def __setitem__(self, key: Incomplete, value: Incomplete) -> None: ... - -no_arg: Incomplete class EnumType(type): @classmethod @@ -273,76 +55,5 @@ class EnumType(type): cls: Incomplete, name: Incomplete, value: Incomplete ) -> None: ... -if StdlibEnumMeta: # type: ignore[truthy-function] - class EnumType(EnumType, StdlibEnumMeta): ... - -EnumMeta = EnumType - -# _reduce_ex_by_name = pickle_by_global_name - -class IntEnum(int, ReprEnum): ... - -class StrEnum(str, ReprEnum): - def __new__( - cls: Incomplete, *values: Incomplete, **kwds: Incomplete - ) -> Incomplete: ... - -class LowerStrEnum(StrEnum): - def __new__( - cls: Incomplete, - value: Incomplete, - *args: Incomplete, - **kwds: Incomplete, - ) -> Incomplete: ... - -class UpperStrEnum(StrEnum): - def __new__( - cls: Incomplete, - value: Incomplete, - *args: Incomplete, - **kwds: Incomplete, - ) -> Incomplete: ... - -class AutoEnum(Enum): ... - -class AutoNumberEnum(Enum): - def __new__( - cls: Incomplete, *args: Incomplete, **kwds: Incomplete - ) -> Incomplete: ... - -class AddValueEnum(Enum): ... -class MultiValueEnum(Enum): ... -class NoAliasEnum(Enum): ... - -class OrderedEnum(Enum): - def __ge__(self, other: Incomplete) -> Incomplete: ... - def __gt__(self, other: Incomplete) -> Incomplete: ... - def __le__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - -class SqliteEnum(Enum): - def __conform__(self, protocol: Incomplete) -> Incomplete: ... - -class UniqueEnum(Enum): ... - -def extend_enum( - enumeration: Incomplete, - name: Incomplete, - *args: Incomplete, - **kwds: Incomplete, -) -> Incomplete: ... -def unique(enumeration: Incomplete) -> Incomplete: ... - -class IntFlag(int, ReprEnum, Flag): - def __contains__(self, other: Incomplete) -> bool: ... - -def add_stdlib_integration() -> None: ... -def remove_stdlib_integration() -> None: ... - -class cls2module: - def __init__(self, cls: Incomplete, *args: Incomplete) -> None: ... - def register(self) -> None: ... - -# Names in __all__ with no definition: -# CONFORM -# STRICT +Enum: EnumType +MultiValue: constant diff --git a/stubs/aenum/_py3.pyi b/stubs/aenum/_py3.pyi deleted file mode 100644 index 7f483ac..0000000 --- a/stubs/aenum/_py3.pyi +++ /dev/null @@ -1,6 +0,0 @@ -__all__ = ["getargspec", "raise_with_traceback", "raise_from_none"] -from _typeshed import Incomplete - -def getargspec(method: Incomplete) -> Incomplete: ... -def raise_with_traceback(exc: Incomplete, tb: Incomplete) -> None: ... -def raise_from_none(exc: Incomplete) -> None: ... diff --git a/stubs/aenum/_tuple.pyi b/stubs/aenum/_tuple.pyi deleted file mode 100644 index 7549f42..0000000 --- a/stubs/aenum/_tuple.pyi +++ /dev/null @@ -1,64 +0,0 @@ -from _typeshed import Incomplete - -from ._common import * -from ._constant import NamedConstant - -__all__ = ["TupleSize", "NamedTuple"] - -class NamedTupleDict(OrderedDict): - def __init__(self, *args: Incomplete, **kwds: Incomplete) -> None: ... - def __setitem__(self, key: Incomplete, value: Incomplete) -> None: ... - -class _TupleAttributeAtIndex: - name: Incomplete - index: Incomplete - __doc__: Incomplete - default: Incomplete - def __init__( - self, - name: Incomplete, - index: Incomplete, - doc: Incomplete, - default: Incomplete, - ) -> None: ... - def __get__( - self, instance: Incomplete, owner: Incomplete - ) -> Incomplete: ... - -class undefined: - def __bool__(self) -> bool: ... - __nonzero__ = __bool__ - -class TupleSize(NamedConstant): - fixed: Incomplete - minimum: Incomplete - variable: Incomplete - -class NamedTupleMeta(type): - @classmethod - def __prepare__( - metacls: Incomplete, - cls: Incomplete, - bases: Incomplete, - size: Incomplete = ..., - **kwds: Incomplete, - ) -> Incomplete: ... - def __init__( - cls: Incomplete, *args: Incomplete, **kwds: Incomplete - ) -> None: ... - def __new__( - metacls: Incomplete, - cls: Incomplete, - bases: Incomplete, - clsdict: Incomplete, - size: Incomplete = ..., - **kwds: Incomplete, - ) -> Incomplete: ... - def __add__(cls: Incomplete, other: Incomplete) -> Incomplete: ... - def __call__( - cls: Incomplete, *args: Incomplete, **kwds: Incomplete - ) -> Incomplete: ... - def __fields__(cls: Incomplete) -> Incomplete: ... - def __aliases__(cls: Incomplete) -> Incomplete: ... - -NamedTuple: Incomplete diff --git a/stubs/aenum/allowlist b/stubs/aenum/allowlist index 1a8b340..396a47d 100644 --- a/stubs/aenum/allowlist +++ b/stubs/aenum/allowlist @@ -1,98 +1,143 @@ -aenum.AutoNumberEnum.__new__ +aenum.AddValue +aenum.AddValueEnum +aenum.AutoEnum +aenum.AutoNumberEnum aenum.CONFORM -aenum.Enum.__ge__ -aenum.Enum.__gt__ -aenum.Enum.__init__ -aenum.Enum.__le__ -aenum.Enum.__lt__ -aenum.Enum._generate_next_value_ -aenum.Enum.values +aenum.Constant +aenum.EJECT aenum.EnumType.__members__ aenum.Flag -aenum.Flag.__and__ -aenum.Flag.__bool__ -aenum.Flag.__contains__ -aenum.Flag.__invert__ -aenum.Flag.__iter__ -aenum.Flag.__len__ -aenum.Flag.__new__ -aenum.Flag.__or__ -aenum.Flag.__rand__ -aenum.Flag.__ror__ -aenum.Flag.__rxor__ -aenum.Flag.__xor__ +aenum.FlagBoundary +aenum.IntEnum aenum.IntFlag -aenum.IntFlag.__new__ -aenum.LowerStrEnum.__new__ -aenum.NamedTuple.__aliases__ -aenum.NamedTuple.__new__ -aenum.NamedTuple.__replace__ -aenum.NamedTuple._fields -aenum.NamedTuple._make +aenum.KEEP +aenum.LowerStrEnum +aenum.MagicValue +aenum.Member +aenum.MultiValueEnum +aenum.NamedConstant +aenum.NamedTuple +aenum.NoAlias +aenum.NoAliasEnum +aenum.NonMember +aenum.OrderedEnum aenum.ReprEnum -aenum.ReprEnum.__new__ aenum.STRICT aenum.SqliteEnum -aenum.StrEnum.__new__ -aenum.TupleSize -aenum.UpperStrEnum.__new__ +aenum.StrEnum +aenum.Unique +aenum.UniqueEnum +aenum.UpperStrEnum aenum.__all__ +aenum._common.Member +aenum._common.NonMember +aenum._common.NoneType +aenum._common.PY2 +aenum._common.PY2_6 +aenum._common.PY3 +aenum._common.PY3_11 +aenum._common.PY3_3 +aenum._common.PY3_4 +aenum._common.PY3_5 +aenum._common.PY3_6 +aenum._common.PY3_7 +aenum._common._Addendum aenum._common.__all__ -aenum._enum.AutoNumberEnum.__new__ +aenum._common._abs_ +aenum._common._add_ +aenum._common._and_ +aenum._common._floordiv_ +aenum._common._inv_ +aenum._common._lshift_ +aenum._common._mod_ +aenum._common._mul_ +aenum._common._neg_ +aenum._common._or_ +aenum._common._pos_ +aenum._common._pow_ +aenum._common._rshift_ +aenum._common._sub_ +aenum._common._truediv_ +aenum._common._value +aenum._common._xor_ +aenum._common.baseinteger +aenum._common.basestring +aenum._common.bltin_property +aenum._common.get_attr_from_chain +aenum._common.getargspec +aenum._common.is_descriptor +aenum._common.is_dunder +aenum._common.is_internal_class +aenum._common.is_private_name +aenum._common.is_sunder +aenum._common.long +aenum._common.make_class_unpicklable +aenum._common.member +aenum._common.nonmember +aenum._common.pyver +aenum._common.raise_from_none +aenum._common.raise_with_traceback +aenum._common.skip +aenum._common.unicode +aenum._enum.AddValue +aenum._enum.AddValueEnum +aenum._enum.AutoEnum +aenum._enum.AutoNumberEnum aenum._enum.CONFORM -aenum._enum.Enum.__ge__ -aenum._enum.Enum.__gt__ -aenum._enum.Enum.__init__ -aenum._enum.Enum.__le__ -aenum._enum.Enum.__lt__ -aenum._enum.Enum._generate_next_value_ -aenum._enum.Enum.values -aenum._enum.EnumConstants -aenum._enum.EnumType-redefinition +aenum._enum.EJECT +aenum._enum.EnumDict +aenum._enum.EnumMeta aenum._enum.EnumType.__members__ aenum._enum.Flag -aenum._enum.Flag.__and__ -aenum._enum.Flag.__bool__ -aenum._enum.Flag.__contains__ -aenum._enum.Flag.__invert__ -aenum._enum.Flag.__iter__ -aenum._enum.Flag.__len__ -aenum._enum.Flag.__new__ -aenum._enum.Flag.__or__ -aenum._enum.Flag.__rand__ -aenum._enum.Flag.__ror__ -aenum._enum.Flag.__rxor__ -aenum._enum.Flag.__xor__ -aenum._enum.FlagBoundary-redefinition -aenum._enum.IntEnum-redefinition +aenum._enum.FlagBoundary +aenum._enum.IntEnum aenum._enum.IntFlag -aenum._enum.IntFlag.__new__ -aenum._enum.LowerStrEnum.__new__ -aenum._enum.RecursionError +aenum._enum.KEEP +aenum._enum.LowerStrEnum +aenum._enum.MagicValue +aenum._enum.MultiValueEnum +aenum._enum.NoAlias +aenum._enum.NoAliasEnum +aenum._enum.OrderedEnum aenum._enum.ReprEnum -aenum._enum.ReprEnum.__new__ aenum._enum.STRICT -aenum._enum.StrEnum.__new__ -aenum._enum.UpperStrEnum.__new__ -aenum._enum._EnumArgSpec +aenum._enum.SqliteEnum +aenum._enum.StrEnum +aenum._enum.Unique +aenum._enum.UniqueEnum +aenum._enum.UpperStrEnum aenum._enum.__all__ aenum._enum._reduce_ex_by_name -aenum._enum.auto.value -aenum._enum.base.deleter -aenum._enum.base.getter -aenum._enum.base.setter -aenum._enum.enum-redefinition +aenum._enum.add_stdlib_integration +aenum._enum.auto +aenum._enum.bin +aenum._enum.bit_count +aenum._enum.bits +aenum._enum.cls2module +aenum._enum.enum_property +aenum._enum.export +aenum._enum.extend_enum aenum._enum.is_single_bit +aenum._enum.no_arg +aenum._enum.property +aenum._enum.remove_stdlib_integration +aenum._enum.show_flag_values +aenum._enum.unique aenum._reduce_ex_by_name -aenum._tuple.NamedTupleMeta.__aliases__ -aenum._tuple.NamedTupleMeta.__fields__ -aenum._tuple.TupleSize -aenum._tuple.undefined -aenum.auto.value +aenum.add_stdlib_integration +aenum.bin +aenum.enum_property +aenum.extend_enum aenum.getargspec -aenum.is_single_bit +aenum.member +aenum.no_arg +aenum.nonmember +aenum.property aenum.raise_from_none aenum.raise_with_traceback +aenum.remove_stdlib_integration +aenum.skip aenum.test aenum.test_v3 -aenum.test_v37 \ No newline at end of file +aenum.test_v37 +aenum.unique \ No newline at end of file diff --git a/stubs/colors/__init__.pyi b/stubs/colors/__init__.pyi index 515bbb3..cf865c1 100644 --- a/stubs/colors/__init__.pyi +++ b/stubs/colors/__init__.pyi @@ -1,3 +1,2 @@ from .colors import * -from .csscolors import * from .version import __version__ as __version__ diff --git a/stubs/colors/allowlist b/stubs/colors/allowlist new file mode 100644 index 0000000..d7fa21d --- /dev/null +++ b/stubs/colors/allowlist @@ -0,0 +1,28 @@ +colors.COLORS +colors.STYLES +colors.colors.COLORS +colors.colors.STYLES +colors.colors.ansilen +colors.colors.black +colors.colors.blink +colors.colors.blink2 +colors.colors.blue +colors.colors.bold +colors.colors.concealed +colors.colors.crossed +colors.colors.cyan +colors.colors.faint +colors.colors.green +colors.colors.is_string +colors.colors.italic +colors.colors.magenta +colors.colors.negative +colors.colors.none +colors.colors.red +colors.colors.string_types +colors.colors.strip_color +colors.colors.underline +colors.colors.white +colors.colors.yellow +colors.css_colors +colors.csscolors \ No newline at end of file diff --git a/stubs/colors/colors.pyi b/stubs/colors/colors.pyi index a16858b..46c8fa9 100644 --- a/stubs/colors/colors.pyi +++ b/stubs/colors/colors.pyi @@ -1,40 +1,10 @@ from __future__ import annotations -from functools import partial -from typing import Any, Final, TypeIs +from typing import Any -from .csscolors import css_colors as css_colors -from .csscolors import parse_rgb as parse_rgb - -string_types: type[str] -COLORS: Final[tuple[str, ...]] -STYLES: Final[tuple[str, ...]] - -def is_string(obj: object) -> TypeIs[str]: ... def color( s: str, fg: str | int | tuple[Any] | None = None, bg: str | int | tuple[Any] | None = None, style: str | None = None, ) -> str: ... -def strip_color(s: str) -> str: ... -def ansilen(s: str) -> int: ... - -black: partial[str] -red: partial[str] -green: partial[str] -yellow: partial[str] -blue: partial[str] -magenta: partial[str] -cyan: partial[str] -white: partial[str] -bold: partial[str] -none: partial[str] -faint: partial[str] -italic: partial[str] -underline: partial[str] -blink: partial[str] -blink2: partial[str] -negative: partial[str] -concealed: partial[str] -crossed: partial[str] diff --git a/stubs/colors/csscolors.pyi b/stubs/colors/csscolors.pyi deleted file mode 100644 index 9574ed2..0000000 --- a/stubs/colors/csscolors.pyi +++ /dev/null @@ -1,8 +0,0 @@ -from __future__ import annotations - -from typing import Final - -string_types: type[str] -css_colors: Final[dict[str, tuple[str, str, str]]] - -def parse_rgb(s: str) -> tuple[str, str, str]: ... diff --git a/stubs/dill/__init__.pyi b/stubs/dill/__init__.pyi index abb72f8..ba445af 100644 --- a/stubs/dill/__init__.pyi +++ b/stubs/dill/__init__.pyi @@ -1,45 +1,6 @@ from __future__ import annotations -from _typeshed import Incomplete from version import __version__ as __version__ # type: ignore -from . import detect as detect -from . import session as session -from . import source as source -from . import temp as temp -from ._dill import CONTENTS_FMODE as CONTENTS_FMODE -from ._dill import DEFAULT_PROTOCOL as DEFAULT_PROTOCOL -from ._dill import FILE_FMODE as FILE_FMODE -from ._dill import HANDLE_FMODE as HANDLE_FMODE -from ._dill import HIGHEST_PROTOCOL as HIGHEST_PROTOCOL -from ._dill import PickleError as PickleError -from ._dill import Pickler as Pickler -from ._dill import PickleWarning as PickleWarning -from ._dill import PicklingError as PicklingError -from ._dill import PicklingWarning as PicklingWarning -from ._dill import Unpickler as Unpickler -from ._dill import UnpicklingError as UnpicklingError -from ._dill import UnpicklingWarning as UnpicklingWarning -from ._dill import check as check -from ._dill import copy as copy -from ._dill import dump as dump from ._dill import dumps as dumps -from ._dill import load as load from ._dill import loads as loads -from ._dill import pickle as pickle -from ._dill import pickles as pickles -from ._dill import register as register -from .session import dump_module as dump_module -from .session import dump_session as dump_session -from .session import load_module as load_module -from .session import load_module_asdict as load_module_asdict -from .session import load_session as load_session -from .settings import settings as settings - -# parent: Incomplete -objects: Incomplete - -def load_types(pickleable: bool = True, unpickleable: bool = True) -> None: ... -def extend(use_dill: bool = True) -> None: ... -def license() -> None: ... -def citation() -> None: ... diff --git a/stubs/dill/_dill.pyi b/stubs/dill/_dill.pyi index 4cc3dca..e32e60b 100644 --- a/stubs/dill/_dill.pyi +++ b/stubs/dill/_dill.pyi @@ -1,94 +1,10 @@ from __future__ import annotations -import typing -from pickle import ( - DEFAULT_PROTOCOL as DEFAULT_PROTOCOL, -) -from pickle import ( - HIGHEST_PROTOCOL as HIGHEST_PROTOCOL, -) -from pickle import ( - PickleError as PickleError, -) -from pickle import ( - PicklingError as PicklingError, -) -from pickle import ( - Unpickler as StockUnpickler, -) -from pickle import ( - UnpicklingError as UnpicklingError, -) -from pickle import ( - _Pickler as StockPickler, -) -from typing import Any, Callable +from pickle import Unpickler as StockUnpickler +from typing import Any from _typeshed import Incomplete -from .logger import adapter as logger - -__all__ = [ - "dump", - "dumps", - "load", - "loads", - "copy", - "Pickler", - "Unpickler", - "register", - "pickle", - "pickles", - "check", - "DEFAULT_PROTOCOL", - "HIGHEST_PROTOCOL", - "HANDLE_FMODE", - "CONTENTS_FMODE", - "FILE_FMODE", - "PickleError", - "PickleWarning", - "PicklingError", - "PicklingWarning", - "UnpicklingError", - "UnpicklingWarning", -] - -log = logger -BufferType = memoryview -ClassType = type -SliceType = slice -TypeType = type -XRangeType = range -try: - IS_IPYTHON = __IPYTHON__ # type: ignore -except NameError: - IS_IPYTHON = False - -class Sentinel: - name: str - __module__: Incomplete - def __init__( - self, name: str, module_name: Incomplete | None = None - ) -> None: ... - def __copy__(self) -> "Sentinel": ... - def __deepcopy__(self, memo: Incomplete) -> "Sentinel": ... - def __reduce__(self) -> str: ... - def __reduce_ex__(self, protocol: Incomplete) -> str: ... - -HANDLE_FMODE: int -CONTENTS_FMODE: int -FILE_FMODE: int - -def copy(obj: object, *args: Any, **kwds: Any) -> Incomplete: ... -def dump( - obj: object, - file: Incomplete, - protocol: Incomplete | None = None, - byref: Incomplete | None = None, - fmode: Incomplete | None = None, - recurse: Incomplete | None = None, - **kwds: Incomplete, -) -> None: ... def dumps( obj: object, protocol: Incomplete | None = None, @@ -97,34 +13,10 @@ def dumps( recurse: Incomplete | None = None, **kwds: Incomplete, ) -> Incomplete: ... -def load( - file: Incomplete, ignore: Incomplete | None = None, **kwds: Any -) -> Unpickler: ... def loads( str: str | bytes, ignore: Incomplete | None = None, **kwds: Any ) -> Unpickler: ... -class MetaCatchingDict(dict): # type: ignore - def get( - self, key: Incomplete, default: Incomplete | None = None - ) -> Incomplete: ... - def __missing__(self, key: Incomplete) -> Incomplete: ... - -class PickleWarning(Warning, PickleError): ... -class PicklingWarning(PickleWarning, PicklingError): ... -class UnpicklingWarning(PickleWarning, UnpicklingError): ... - -class Pickler(StockPickler): - dispatch: dict[type, Callable[[Pickler, typing.Any], None]] # type: ignore - settings: dict[str, Any] - def __init__( - self, file: Incomplete, *args: Incomplete, **kwds: Incomplete - ) -> None: ... - def save( - self, obj: Incomplete, save_persistent_id: bool = True - ) -> None: ... - def dump(self, obj: Incomplete) -> None: ... - class Unpickler(StockUnpickler): settings: dict[str, Any] def find_class( @@ -132,38 +24,3 @@ class Unpickler(StockUnpickler): ) -> Incomplete: ... def __init__(self, *args: Incomplete, **kwds: Incomplete) -> None: ... def load(self) -> Incomplete: ... - -def pickle(t: Incomplete, func: Incomplete) -> None: ... -def register(t: Incomplete) -> Callable: ... # type: ignore - -class match: - value: Incomplete - def __init__(self, value: Incomplete) -> None: ... - def __enter__(self) -> "match": ... - def __exit__(self, *exc_info: Incomplete) -> bool: ... - args: Incomplete - def case(self, args: Any) -> bool: ... - @property - def fields(self) -> dict: ... # type: ignore - def __getattr__(self, item: Any) -> Any: ... - -CODE_VERSION: tuple[int, ...] - -# class _itemgetter_helper: -# items: Incomplete -# def __init__(self) -> None: ... -# def __getitem__(self, item) -> None: ... -# -# class _attrgetter_helper: -# attrs: Incomplete -# index: Incomplete -# def __init__(self, attrs, index: Incomplete | None = None) -> None: ... -# def __getattribute__(self, attr): ... -# -# class _dictproxy_helper(dict): -# def __ror__(self, a): ... - -def pickles( - obj: object, exact: bool = False, safe: bool = False, **kwds: Any -) -> bool | Incomplete: ... -def check(obj: object, *args: Any, **kwds: Any) -> None: ... diff --git a/stubs/dill/_objects.pyi b/stubs/dill/_objects.pyi deleted file mode 100644 index 04aba41..0000000 --- a/stubs/dill/_objects.pyi +++ /dev/null @@ -1,26 +0,0 @@ -from __future__ import annotations - -import ctypes - -from _typeshed import Incomplete - -__all__ = ["registered", "failures", "succeeds"] - -class _class: ... - -class _class2: - def __call__(self) -> None: ... - -class _newclass: ... - -class _newclass2: - descriptor: Incomplete - -class _Struct(ctypes.Structure): - next: Incomplete - -registered: Incomplete - -failures: Incomplete - -succeeds: Incomplete diff --git a/stubs/dill/_shims.pyi b/stubs/dill/_shims.pyi deleted file mode 100644 index 0839f82..0000000 --- a/stubs/dill/_shims.pyi +++ /dev/null @@ -1,28 +0,0 @@ -from __future__ import annotations - -from typing import Any, Callable - -from _typeshed import Incomplete - -class Reduce: - reduction: tuple[Any, ...] - def __new__( - cls: object, *reduction: tuple[Any, ...], **kwargs: dict[str, bool] - ) -> "Reduce": ... - def __copy__(self) -> "Reduce": ... - def __deepcopy__(self, memo: Any) -> "Reduce": ... - def __reduce__(self) -> tuple[Any, ...]: ... - def __reduce_ex__(self, protocol: Any) -> tuple[Any, ...]: ... - -class _CallableReduce(Reduce): - def __call__( - self, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - -def Getattr( - object: object, name: str, default: Incomplete = ... -) -> Reduce: ... -def move_to( - module: Incomplete, name: Incomplete | None = None -) -> Callable: ... # type: ignore -def register_shim(name: Incomplete, default: Incomplete) -> Reduce: ... diff --git a/stubs/dill/allowlist b/stubs/dill/allowlist index 3ba93fd..29c792b 100644 --- a/stubs/dill/allowlist +++ b/stubs/dill/allowlist @@ -1,3 +1,31 @@ +dill._dill.CONTENTS_FMODE +dill._dill.FILE_FMODE +dill._dill.HANDLE_FMODE +dill._dill.PickleWarning +dill._dill.Pickler +dill._dill.PicklingWarning +dill._dill.UnpicklingWarning +dill._dill.__all__ +dill._dill.check +dill._dill.copy +dill._dill.dump +dill._dill.load +dill._dill.pickle +dill._dill.pickles +dill._dill.register +dill.citation +dill.detect +dill.extend +dill.license +dill.load_types +dill.logger +dill.objects +dill.objtypes +dill.pointers +dill.session +dill.settings +dill.source +dill.temp dill.tests dill.tests.__main__ dill.tests.test_abc @@ -29,5 +57,4 @@ dill.tests.test_source dill.tests.test_sources dill.tests.test_temp dill.tests.test_threads -dill.tests.test_weakref -dill.__diff.pyi \ No newline at end of file +dill.tests.test_weakref \ No newline at end of file diff --git a/stubs/dill/detect.pyi b/stubs/dill/detect.pyi deleted file mode 100644 index ee68edb..0000000 --- a/stubs/dill/detect.pyi +++ /dev/null @@ -1,56 +0,0 @@ -from __future__ import annotations - -from typing import Callable, Dict, MappingView, Type - -from _typeshed import Incomplete - -from .logger import trace as trace - -__all__ = [ - "baditems", - "badobjects", - "badtypes", - "code", - "errors", - "freevars", - "getmodule", - "globalvars", - "nestedcode", - "nestedglobals", - "outermost", - "referredglobals", - "referrednested", - "trace", - "varnames", -] - -def getmodule( - object: object, _filename: Incomplete | None = None, force: bool = False -) -> Incomplete: ... -def outermost(func: Callable) -> Incomplete: ... # type: ignore -def nestedcode(func: Callable, recurse: bool = True) -> list: ... # type: ignore -def code(func: Callable) -> Incomplete: ... # type: ignore -def referrednested(func: Callable, recurse: bool = True) -> list: ... # type: ignore -def freevars(func: Callable) -> dict: ... # type: ignore -def nestedglobals(func: Callable, recurse: bool = True) -> list: ... # type: ignore -def referredglobals( - func: Callable, # type: ignore - recurse: bool = True, - builtin: bool = False, -) -> MappingView: ... -def globalvars( - func: Callable, # type: ignore - recurse: bool = ..., - builtin: bool = ..., -) -> Dict[str, Type[KeyError]]: ... -def varnames(func: Callable) -> tuple: ... # type: ignore -def baditems(obj: object, exact: bool = False, safe: bool = False) -> list: ... # type: ignore -def badobjects( - obj: object, depth: int = 0, exact: bool = False, safe: bool = False -) -> dict: ... # type: ignore -def badtypes( - obj: object, depth: int = 0, exact: bool = False, safe: bool = False -) -> dict: ... # type: ignore -def errors( - obj: object, depth: int = 0, exact: bool = False, safe: bool = False -) -> Incomplete: ... diff --git a/stubs/dill/logger.pyi b/stubs/dill/logger.pyi deleted file mode 100644 index 8831ea9..0000000 --- a/stubs/dill/logger.pyi +++ /dev/null @@ -1,58 +0,0 @@ -from __future__ import annotations - -import contextlib -import logging -from logging import ( - Logger, - StreamHandler, -) -from os import PathLike -from typing import ( - Any, - Optional, - TextIO, - Union, -) - -from _typeshed import Incomplete - -from dill._dill import Pickler - -__all__ = ["adapter", "logger", "trace"] - -class TraceAdapter: - def __init__(self, logger: Logger) -> None: ... - def addHandler(self, handler: StreamHandler) -> None: ... # type: ignore - def trace( - self, pickler: Pickler, msg: str, *args: Any, **kwargs: Any - ) -> None | TraceManager: ... - def trace_setup(self, pickler: Pickler) -> None: ... - def process(self, msg: Any, kwargs: Any) -> tuple[Any, Any]: ... - def removeHandler(self, handler: Any) -> None: ... - -class TraceFormatter(logging.Formatter): - is_utf8: bool - def __init__( - self, *args: Any, handler: Incomplete = ..., **kwargs: Any - ): ... - def format(self, record: Incomplete) -> Incomplete: ... - -logger: Incomplete -adapter: Incomplete - -def trace( - arg: Optional[Union[bool, TextIO, str, PathLike]] = ..., # type: ignore - *, - mode: str = ..., -) -> None: ... - -class TraceManager(contextlib.AbstractContextManager): # type: ignore - file: Any - mode: Any - redirect: bool - file_is_stream: bool - def __init__(self, file: Any, mode: Any) -> None: ... - handler: logging.StreamHandler | logging.FileHandler # type: ignore - old_level: Incomplete - def __enter__(self) -> Incomplete: ... - def __exit__(self, *exc_info: Incomplete) -> None: ... diff --git a/stubs/dill/objtypes.pyi b/stubs/dill/objtypes.pyi deleted file mode 100644 index 8bcfc1f..0000000 --- a/stubs/dill/objtypes.pyi +++ /dev/null @@ -1 +0,0 @@ -# from dill import objects as objects diff --git a/stubs/dill/pointers.pyi b/stubs/dill/pointers.pyi deleted file mode 100644 index 8c61940..0000000 --- a/stubs/dill/pointers.pyi +++ /dev/null @@ -1,24 +0,0 @@ -from __future__ import annotations - -from typing import Any - -from ._dill import _locate_object as at # type: ignore -from ._dill import _proxy_helper as reference # type: ignore - -__all__ = ["parent", "reference", "at", "parents", "children"] - -def parent(obj: object, objtype: type, ignore: tuple = ()) -> Any: ... # type: ignore -def parents( - obj: object, - objtype: type, - depth: int = 1, - ignore: tuple = (), # type: ignore -) -> Any: ... -def children( - obj: object, - objtype: type, - depth: int = 1, - ignore: tuple = (), # type: ignore -) -> Any: ... - -refobject = at diff --git a/stubs/dill/session.pyi b/stubs/dill/session.pyi deleted file mode 100644 index b8e25df..0000000 --- a/stubs/dill/session.pyi +++ /dev/null @@ -1,53 +0,0 @@ -from __future__ import annotations - -import os - -from _typeshed import Incomplete - -from ._dill import ModuleType # type: ignore - -__all__ = [ - "dump_module", - "load_module", - "load_module_asdict", - "dump_session", - "load_session", -] - -def dump_module( - filename: str | os.PathLike | None = None, # type: ignore - module: ModuleType | str | None = None, - refimported: bool = False, - **kwds: Incomplete, -) -> None: ... -def dump_session( - filename: Incomplete | None = None, - main: Incomplete | None = None, - byref: bool = False, - **kwds: Incomplete, -) -> None: ... - -class _PeekableReader: - stream: Incomplete - def __init__(self, stream: Incomplete) -> None: ... - def read(self, n: Incomplete) -> Incomplete: ... - def readline(self) -> Incomplete: ... - def tell(self) -> Incomplete: ... - def close(self) -> Incomplete: ... - def peek(self, n: Incomplete) -> Incomplete: ... - -def load_module( - filename: str | os.PathLike | None = None, # type: ignore - module: ModuleType | str | None = None, - **kwds: Incomplete, -) -> ModuleType | None: ... -def load_session( - filename: Incomplete | None = None, - main: Incomplete | None = None, - **kwds: Incomplete, -) -> None: ... -def load_module_asdict( - filename: str | os.PathLike | None = None, # type: ignore - update: bool = False, - **kwds: Incomplete, -) -> dict: ... # type: ignore diff --git a/stubs/dill/settings.pyi b/stubs/dill/settings.pyi deleted file mode 100644 index d208c30..0000000 --- a/stubs/dill/settings.pyi +++ /dev/null @@ -1,5 +0,0 @@ -from __future__ import annotations - -from _typeshed import Incomplete - -settings: Incomplete diff --git a/stubs/dill/source.pyi b/stubs/dill/source.pyi deleted file mode 100644 index bb829c3..0000000 --- a/stubs/dill/source.pyi +++ /dev/null @@ -1,71 +0,0 @@ -from __future__ import annotations - -from typing import Any, Callable, List, Tuple - -from _typeshed import Incomplete - -__all__ = [ - "findsource", - "getsourcelines", - "getsource", - "indent", - "outdent", - "_wrap", - "dumpsource", - "getname", - "_namespace", - "getimport", - "_importable", - "importable", - "isdynamic", - "isfrommain", -] - -def isfrommain(obj: Any) -> bool: ... -def isdynamic(obj: Any) -> bool: ... -def findsource(object: Any) -> Tuple[List[str], int]: ... -def getsourcelines( - object: Any, lstrip: bool = False, enclosing: bool = False -) -> Tuple[List[str], int]: ... -def getsource( - object: Any, - alias: str = "", - lstrip: bool = False, - enclosing: bool = False, - force: bool = False, - builtin: bool = False, -) -> str: ... -def indent(code: str, spaces: int = 4) -> str: ... -def outdent( - code: str, spaces: Incomplete | None = None, all: bool = True -) -> str: ... -def _wrap(f: Callable) -> Callable: ... # type: ignore -def dumpsource( - object: Any, alias: str = "", new: bool = False, enclose: bool = True -) -> str: ... -def getname(obj: Any, force: bool = False, fqn: bool = False) -> str: ... -def _namespace(obj: Any) -> str: ... -def getimport( - obj: Any, - alias: str = "", - verify: bool = True, - builtin: bool = False, - enclosing: bool = False, -) -> str: ... -def _importable( - obj: Any, - alias: str = "", - source: Incomplete | None = None, - enclosing: bool = False, - force: bool = True, - builtin: bool = True, - lstrip: bool = True, -) -> str: ... -def importable( - obj: Any, - alias: str = "", - source: Incomplete | None = None, - builtin: bool = True, -) -> str: ... - -# getblocks_from_history = getblocks diff --git a/stubs/dill/temp.pyi b/stubs/dill/temp.pyi deleted file mode 100644 index 3a7baf4..0000000 --- a/stubs/dill/temp.pyi +++ /dev/null @@ -1,26 +0,0 @@ -from __future__ import annotations - -from io import StringIO -from typing import IO, Any, Iterator - -__all__ = [ - "dump_source", - "dump", - "dumpIO_source", - "dumpIO", - "load_source", - "load", - "loadIO_source", - "loadIO", - "capture", -] - -def capture(stream: str = ...) -> Iterator[StringIO]: ... -def load_source(file: str, **kwds: Any) -> Any: ... -def dump_source(object: Any, **kwds: Any) -> None: ... -def load(file: str | IO[bytes], **kwds: Any) -> Any: ... -def dump(object: Any, **kwds: Any) -> None: ... -def loadIO(buffer: IO[bytes], **kwds: Any) -> Any: ... -def dumpIO(object: Any, **kwds: Any) -> None: ... -def loadIO_source(buffer: IO[bytes], **kwds: Any) -> Any: ... -def dumpIO_source(object: Any, **kwds: Any) -> None: ... diff --git a/stubs/lemminflect/__init__.pyi b/stubs/lemminflect/__init__.pyi index 95fb0e8..f8ea0d1 100644 --- a/stubs/lemminflect/__init__.pyi +++ b/stubs/lemminflect/__init__.pyi @@ -1,24 +1,6 @@ -# Auto-generated by monkeytype - -from typing import Dict, Optional, Tuple - -def getAllInflections( - lemma: str, upos: Optional[str] = None -) -> Dict[str, Tuple[str, ...]]: ... -def getAllInflectionsOOV( - lemma: str, upos: str -) -> Dict[str, Tuple[str, ...]]: ... -def getAllLemmas( - word: str, upos: Optional[str] = None -) -> Dict[str, Tuple[str, ...]]: ... -def getAllLemmasOOV( - word: str, upos: Optional[str] -) -> dict[str, tuple[str, ...]]: ... def getInflection( - lemma: str, tag: str, inflect_oov: bool = ... -) -> Tuple[str, ...]: ... + lemma: str, tag: str, inflect_oov: bool = True +) -> tuple[str, ...]: ... def getLemma( - word: str, upos: str, lemmatize_oov: bool = ... -) -> Tuple[str, ...]: ... -def isTagBaseForm(tag: str) -> bool: ... -def setUseInternalLemmatizer(TF: bool) -> None: ... + word: str, upos: str, lemmatize_oov: bool = True +) -> tuple[str, ...]: ... diff --git a/stubs/lemminflect/allowlist b/stubs/lemminflect/allowlist index db2ae08..f23dafe 100644 --- a/stubs/lemminflect/allowlist +++ b/stubs/lemminflect/allowlist @@ -1,7 +1,38 @@ -lemminflect.core.Inflections.Inflections -lemminflect.core.Inflections.Inflections.spacyGetInfl -lemminflect.core.Lemmatizer.Lemmatizer -lemminflect.core.Lemmatizer.Lemmatizer.spacyGetLemma -lemminflect.kmodels.KInfer.KInferWithKeras -lemminflect.kmodels.KInfer.KInferWithNumpy -lemminflect.utils.Singleton.Singleton +lemminflect.codecs +lemminflect.codecs.FormsTable +lemminflect.codecs.InflTCorpFileCodec +lemminflect.codecs.InflectionLUCodec +lemminflect.codecs.LemmaLUCodec +lemminflect.codecs.LemmaTCorpFileCodec +lemminflect.codecs.OverridesCodec +lemminflect.config +lemminflect.core +lemminflect.core.InflectionRules +lemminflect.core.Inflections +lemminflect.core.Lemmatizer +lemminflect.core.LemmatizerRules +lemminflect.core.LexicalUtils +lemminflect.getAllInflections +lemminflect.getAllInflectionsOOV +lemminflect.getAllLemmas +lemminflect.getAllLemmasOOV +lemminflect.isTagBaseForm +lemminflect.kmodels +lemminflect.kmodels.KInfer +lemminflect.kmodels.KerasModel +lemminflect.kmodels.ModelInfl +lemminflect.kmodels.ModelInflInData +lemminflect.kmodels.ModelLemma +lemminflect.kmodels.ModelLemmaClasses +lemminflect.kmodels.ModelLemmaInData +lemminflect.setUseInternalLemmatizer +lemminflect.slexicon +lemminflect.slexicon.SKey +lemminflect.slexicon.SPECIALISTEntry +lemminflect.slexicon.SPECIALISTExtractor +lemminflect.utils +lemminflect.utils.CorpusUtils +lemminflect.utils.DataContainer +lemminflect.utils.ProgressBar +lemminflect.utils.Singleton +lemminflect.utils.Unigrams \ No newline at end of file diff --git a/stubs/lemminflect/codecs/FormsTable.pyi b/stubs/lemminflect/codecs/FormsTable.pyi deleted file mode 100644 index 89dc836..0000000 --- a/stubs/lemminflect/codecs/FormsTable.pyi +++ /dev/null @@ -1,7 +0,0 @@ -# Auto-generated by stubgen - -from _typeshed import Incomplete - -class FormsTable: - data: Incomplete - def __init__(self, table_fn: Incomplete) -> None: ... diff --git a/stubs/lemminflect/codecs/InflTCorpFileCodec.pyi b/stubs/lemminflect/codecs/InflTCorpFileCodec.pyi deleted file mode 100644 index e6e5da7..0000000 --- a/stubs/lemminflect/codecs/InflTCorpFileCodec.pyi +++ /dev/null @@ -1,21 +0,0 @@ -# Auto-generated by stubgen - -from _typeshed import Incomplete - -class Entry: - lemma: Incomplete - category: Incomplete - source: Incomplete - def __init__( - self, lemma: Incomplete, category: Incomplete, source: Incomplete - ) -> None: ... - -class InflTCorpFileCodec: - @staticmethod - def fromString(line: Incomplete) -> Incomplete: ... - @staticmethod - def toString( - lemma: Incomplete, category: Incomplete, source: Incomplete - ) -> Incomplete: ... - @classmethod - def load(cls: Incomplete, fn: Incomplete) -> Incomplete: ... diff --git a/stubs/lemminflect/codecs/InflectionLUCodec.pyi b/stubs/lemminflect/codecs/InflectionLUCodec.pyi deleted file mode 100644 index 1f4e065..0000000 --- a/stubs/lemminflect/codecs/InflectionLUCodec.pyi +++ /dev/null @@ -1,40 +0,0 @@ -# Auto-generated by monkeytype - -from typing import ( - Any, - Dict, - Tuple, -) - -class InflectionLUCodec: - penn_dict: dict[str, list[str]] - slex_dict: dict[str, list[str]] - @classmethod - def fromString(cls, line: str) -> Any: ... - @classmethod - def load( - cls, fn: str - ) -> Dict[ - str, - Dict[ - str, - Tuple[str, ...], - ], - ]: ... - @staticmethod - def updateForAuxMod( - d: Dict[ - str, - Dict[str, Tuple[str, ...]], - ], - ) -> Dict[ - str, - Dict[str, Tuple[str, ...]], - ]: ... - @classmethod - def toString( - cls: object, - word: str, - category: str, - forms_dict: dict[str, str], - ) -> str: ... diff --git a/stubs/lemminflect/codecs/LemmaLUCodec.pyi b/stubs/lemminflect/codecs/LemmaLUCodec.pyi deleted file mode 100644 index 3c1055c..0000000 --- a/stubs/lemminflect/codecs/LemmaLUCodec.pyi +++ /dev/null @@ -1,22 +0,0 @@ -# Auto-generated by monkeytype - -from typing import Dict, Iterable, Tuple, Union - -class LemmaLUCodec: - @staticmethod - def fromString( - line: str, - ) -> Union[ - Tuple[str, str, Tuple[str, str, str]], - Tuple[str, str, Tuple[str]], - Tuple[str, str, Tuple[str, str]], - ]: ... - @classmethod - def load( - cls, fn: str - ) -> Dict[ - str, - Dict[str, Union[Tuple[str, str], Tuple[str], Tuple[str, str, str]]], - ]: ... - @staticmethod - def toString(word: str, category: str, forms: Iterable) -> str: ... # type: ignore diff --git a/stubs/lemminflect/codecs/LemmaTCorpFileCodec.pyi b/stubs/lemminflect/codecs/LemmaTCorpFileCodec.pyi deleted file mode 100644 index 7d17ef6..0000000 --- a/stubs/lemminflect/codecs/LemmaTCorpFileCodec.pyi +++ /dev/null @@ -1,29 +0,0 @@ -# Auto-generated by stubgen - -from _typeshed import Incomplete - -class Entry: - infl: Incomplete - category: Incomplete - source: Incomplete - lemma: Incomplete - def __init__( - self, - infl: Incomplete, - category: Incomplete, - source: Incomplete, - lemma: Incomplete, - ) -> None: ... - -class LemmaTCorpFileCodec: - @staticmethod - def fromString(line: Incomplete) -> Incomplete: ... - @staticmethod - def toString( - infl: Incomplete, - category: Incomplete, - source: Incomplete, - lemma: Incomplete, - ) -> Incomplete: ... - @classmethod - def load(cls: Incomplete, fn: Incomplete) -> Incomplete: ... diff --git a/stubs/lemminflect/codecs/OverridesCodec.pyi b/stubs/lemminflect/codecs/OverridesCodec.pyi deleted file mode 100644 index d5c5c92..0000000 --- a/stubs/lemminflect/codecs/OverridesCodec.pyi +++ /dev/null @@ -1,12 +0,0 @@ -# Auto-generated by monkeytype - -from typing import ( - Dict, - Tuple, -) - -class OverridesCodec: - @staticmethod - def load(fn: str) -> Dict[str, Dict[str, Tuple[str]]]: ... - @staticmethod - def toString(word_in: str, pos: str, word_out: str) -> str: ... diff --git a/stubs/lemminflect/codecs/__init__.pyi b/stubs/lemminflect/codecs/__init__.pyi deleted file mode 100644 index f68e791..0000000 --- a/stubs/lemminflect/codecs/__init__.pyi +++ /dev/null @@ -1 +0,0 @@ -# Auto-generated by stubgen diff --git a/stubs/lemminflect/config.pyi b/stubs/lemminflect/config.pyi deleted file mode 100644 index d91fcea..0000000 --- a/stubs/lemminflect/config.pyi +++ /dev/null @@ -1,26 +0,0 @@ -# Auto-generated by stubgen - -data_repo: str -proj_resources: str -gutenberg_dir: str -bwcorp_dir: str -unigrams_gb_all_fn: str -unigrams_gb_clean_fn: str -unigrams_bw_all_fn: str -unigrams_bw_clean_fn: str -unigrams_fn: str -english_dict_fn: str -lexicon_fn: str -ftable_fn: str -lemma_lu_fn: str -inflection_lu_fn: str -model_lemma_cl_fn: str -lemma_tcorp_fn: str -infl_tcorp_fn: str -model_lemma_fn: str -model_infl_fn: str -lemma_overrides_fn: str -infl_overrides_fn: str -acc_word_set_fn: str -acc_lemma_corp_fn: str -kinfer_type: str diff --git a/stubs/lemminflect/core/InflectionRules.pyi b/stubs/lemminflect/core/InflectionRules.pyi deleted file mode 100644 index 0c421b3..0000000 --- a/stubs/lemminflect/core/InflectionRules.pyi +++ /dev/null @@ -1,26 +0,0 @@ -# Auto-generated by monkeytype - -from typing import ( - Dict, - List, -) - -class InflectionRules: - @classmethod - def buildDoubledAdjAdv(cls, base: str) -> List[str]: ... - @classmethod - def buildDoubledVerb(cls, base: str) -> List[str]: ... - @classmethod - def buildGrecNoun(cls, base: str) -> List[str]: ... - @classmethod - def buildRegAdjAdv(cls, base: str) -> List[str]: ... - @classmethod - def buildRegNoun(cls, base: str) -> List[str]: ... - @classmethod - def buildRegVerb(cls, base: str) -> List[str]: ... - @classmethod - def morph(cls, base: str, upos: str, infl_type: str) -> Dict[str, str]: ... - -class MorphologyStyleModel: - def __init__(self, kitype: str = ..., model_fn: str = ...): ... - def getStyle(self, lemma: str, upos: str) -> str: ... diff --git a/stubs/lemminflect/core/Inflections.pyi b/stubs/lemminflect/core/Inflections.pyi deleted file mode 100644 index 2a66bac..0000000 --- a/stubs/lemminflect/core/Inflections.pyi +++ /dev/null @@ -1,48 +0,0 @@ -# Auto-generated by monkeytype - -from typing import ( - Dict, - Optional, - Tuple, -) - -from lemminflect.core.InflectionRules import MorphologyStyleModel - -from ..utils.Singleton import _Singleton - -# from spacy.tokens.token import Token - -class Inflections(_Singleton): - DICT_UPOS_TYPES: list[str] - def __init__(self, infl_lu_fn: str = ..., overrides_fn: str = ...): ... - @staticmethod - def _extractForm( - forms: Dict[str, Tuple[str, ...]], tag: str - ) -> Optional[Tuple[str, ...]]: ... - def _getInflDict( - self, - ) -> Dict[ - str, - Dict[str, Tuple[str, ...]], - ]: ... - def _getInflStyleModel(self) -> MorphologyStyleModel: ... - def _getOverridesDict(self) -> Dict[str, Dict[str, Tuple[str]]]: ... - def getAllInflections( - self, lemma: str, upos: Optional[str] = ... - ) -> Dict[str, Tuple[str, ...]]: ... - def getAllInflectionsOOV( - self, lemma: str, upos: Optional[str] - ) -> Dict[str, Tuple[str, ...]]: ... - def getInflection( - self, lemma: str, tag: str, inflect_oov: bool = ... - ) -> Tuple[str, ...]: ... - def setUseInternalLemmatizer(self, TF: bool) -> None: ... - # def spacyGetInfl( - # self, - # token: Token, - # tag: str, - # form_num: int = ..., - # inflect_oov: bool = ..., - # on_empty_ret_word: bool = ..., - # ) -> str: ... - def isUsingInternalLemmatizer(self) -> bool: ... diff --git a/stubs/lemminflect/core/Lemmatizer.pyi b/stubs/lemminflect/core/Lemmatizer.pyi deleted file mode 100644 index 42a66fa..0000000 --- a/stubs/lemminflect/core/Lemmatizer.pyi +++ /dev/null @@ -1,43 +0,0 @@ -# Auto-generated by monkeytype - -from typing import ( - Dict, - Optional, - Tuple, -) - -from lemminflect.core.LemmatizerRules import LemmatizerRules - -from ..utils.Singleton import _Singleton - -# from spacy.tokens.token import Token - -class Lemmatizer(_Singleton): - DICT_UPOS_TYPES: list[str] - def __init__(self, lemma_lu_fn: str = ..., overrides_fn: str = ...): ... - def _getLemmaDict( - self, - ) -> Dict[ - str, - Dict[str, Tuple[str, ...]], - ]: ... - def _getOOVLemmatizer(self) -> LemmatizerRules: ... - def _getOverridesDict(self) -> Dict[str, Dict[str, Tuple[str]]]: ... - def getAllLemmas( - self, word: str, upos: Optional[str] = ... - ) -> Dict[str, Tuple[str, ...]]: ... - def getAllLemmasOOV( - self, word: str, upos: str - ) -> Dict[str, Tuple[str]]: ... - def getLemma( - self, word: str, upos: str, lemmatize_oov: bool = ... - ) -> Tuple[str, ...]: ... - @staticmethod - def isTagBaseForm(tag: str) -> bool: ... - # def spacyGetLemma( - # self, - # token: Token, - # form_num: int = ..., - # lemmatize_oov: bool = ..., - # on_empty_ret_word: bool = ..., - # ) -> str: ... diff --git a/stubs/lemminflect/core/LemmatizerRules.pyi b/stubs/lemminflect/core/LemmatizerRules.pyi deleted file mode 100644 index 2a0b8a1..0000000 --- a/stubs/lemminflect/core/LemmatizerRules.pyi +++ /dev/null @@ -1,12 +0,0 @@ -# Auto-generated by monkeytype - -from typing import List - -from numpy import int64 - -class LemmatizerRules: - def __init__(self, kitype: str = ..., model_fn: str = ...): ... - def _applyRule(self, inflection: str, rnum: int64) -> str: ... - @staticmethod - def _csvToTuple(line: str) -> List[str]: ... - def lemmatize(self, word: str, upos: str) -> str: ... diff --git a/stubs/lemminflect/core/LexicalUtils.pyi b/stubs/lemminflect/core/LexicalUtils.pyi deleted file mode 100644 index 54639b5..0000000 --- a/stubs/lemminflect/core/LexicalUtils.pyi +++ /dev/null @@ -1,19 +0,0 @@ -# Auto-generated by monkeytype - -from typing import Any, Dict, List, Tuple, Union - -def applyCapsStyle(word: str, style: str) -> str: ... -def applyCapsStyleToDict( - data: Dict[str, Tuple[str, ...]], style: str -) -> Dict[str, Tuple[str, ...]]: ... -def categoryToUPos(category: str) -> str: ... -def getCapsStyle(word: str) -> str: ... -def pennTagAlts(tag: str) -> List[Union[Any, str]]: ... -def tagToUPos(tag: str) -> str | None: ... -def uposToCategory(upos: str) -> str: ... -def uposToTags(upos: str) -> List[str]: ... -def catAndTypeToTag( - category: str, infl_type: str, is_proper: bool = False -) -> list[str]: ... - -slex_dict: dict[tuple[str, str, bool] | tuple[str, str], list[str]] diff --git a/stubs/lemminflect/core/__init__.pyi b/stubs/lemminflect/core/__init__.pyi deleted file mode 100644 index f68e791..0000000 --- a/stubs/lemminflect/core/__init__.pyi +++ /dev/null @@ -1 +0,0 @@ -# Auto-generated by stubgen diff --git a/stubs/lemminflect/kmodels/KInfer.pyi b/stubs/lemminflect/kmodels/KInfer.pyi deleted file mode 100644 index 5c97b53..0000000 --- a/stubs/lemminflect/kmodels/KInfer.pyi +++ /dev/null @@ -1,38 +0,0 @@ -# Auto-generated by monkeytype - -from abc import ABC, abstractmethod -from typing import Any, Dict, List, Tuple, Union - -from _typeshed import Incomplete -from numpy import ( - int64, - ndarray, -) - -def applyActivation(x: ndarray[Any, Any], atype: str) -> ndarray[Any, Any]: ... -def flatten(x: ndarray[Any, Any]) -> ndarray[Any, Any]: ... -def getKInferInstance(kitype: str, model_fn: str) -> KInferWithNumpy: ... -def relu(x: ndarray[Any, Any]) -> ndarray[Any, Any]: ... -def softmax(x: ndarray[Any, Any], axis: int = ...) -> ndarray[Any, Any]: ... - -class KInfer(ABC): - def __init__(self) -> None: ... - def _loadModelContainer( - self, fn: str - ) -> Tuple[ - Dict[str, Union[str, List[Any], Dict[str, Any]]], - List[ndarray[Any, Any]], - ]: ... - def _netOutToValue(self, vec: ndarray[Any, Any]) -> Tuple[int64, str]: ... - def getOutputEnum(self) -> List[str]: ... - @abstractmethod - def run(self, in_vec: ndarray[Any, Any]) -> Incomplete: ... - -class KInferWithKeras: - def __init__(self, fn: str): ... - def _load(self, fn: str): ... # type: ignore - def run(self, in_vec: ndarray[Any, Any]) -> Incomplete: ... - -class KInferWithNumpy: - def __init__(self, fn: str): ... - def run(self, in_vec: ndarray[Any, Any]) -> Tuple[int64, str]: ... diff --git a/stubs/lemminflect/kmodels/KerasModel.pyi b/stubs/lemminflect/kmodels/KerasModel.pyi deleted file mode 100644 index 5c2b2db..0000000 --- a/stubs/lemminflect/kmodels/KerasModel.pyi +++ /dev/null @@ -1,15 +0,0 @@ -# Auto-generated by monkeytype - -from typing import Any, List - -def limitTFMem(pct_mem_res: None = ...): ... # type: ignore - -class KerasModel: - def __init__(self) -> None: ... - @classmethod - def load(cls, filename: str) -> "KerasModel": ... - def save(self, filename: str) -> None: ... - def getInputShape(self) -> Any: ... - def run( - self, inputMat3D: List[List[List[float]]] - ) -> List[List[float]]: ... diff --git a/stubs/lemminflect/kmodels/ModelInfl.pyi b/stubs/lemminflect/kmodels/ModelInfl.pyi deleted file mode 100644 index 740f6f3..0000000 --- a/stubs/lemminflect/kmodels/ModelInfl.pyi +++ /dev/null @@ -1,23 +0,0 @@ -# Auto-generated by stubgen - -from _typeshed import Incomplete - -from .KerasModel import KerasModel - -class ModelInfl(KerasModel): - def __init__(self) -> None: ... - model: Incomplete - def create( - self, - input_len: Incomplete, - input_letters: Incomplete, - output_classes: Incomplete, - ) -> None: ... - def train( - self, - xdata: Incomplete, - ydata: Incomplete, - batch_size: Incomplete, - nepochs: Incomplete, - **kwargs: Incomplete, - ) -> None: ... diff --git a/stubs/lemminflect/kmodels/ModelInflInData.pyi b/stubs/lemminflect/kmodels/ModelInflInData.pyi deleted file mode 100644 index 41e0ad4..0000000 --- a/stubs/lemminflect/kmodels/ModelInflInData.pyi +++ /dev/null @@ -1,14 +0,0 @@ -# Auto-generated by stubgen - -from typing import Any - -from numpy import ndarray - -class ModelInflInData: - MAX_LETTER_IDX: int - WVEC_LEN: int - def __init__(self, fn: str): ... - @classmethod - def wordToVec(cls, word: str, category: str) -> ndarray[Any, Any]: ... - @staticmethod - def getLetterClasses() -> list[str]: ... diff --git a/stubs/lemminflect/kmodels/ModelLemma.pyi b/stubs/lemminflect/kmodels/ModelLemma.pyi deleted file mode 100644 index 6d7d2c7..0000000 --- a/stubs/lemminflect/kmodels/ModelLemma.pyi +++ /dev/null @@ -1,24 +0,0 @@ -# Auto-generated by stubgen - -from _typeshed import Incomplete - -from .KerasModel import KerasModel as KerasModel -from .KerasModel import limitTFMem as limitTFMem - -class ModelLemma(KerasModel): - def __init__(self) -> None: ... - model: Incomplete - def create( - self, - input_len: Incomplete, - input_letters: Incomplete, - output_rules: Incomplete, - ) -> None: ... - def train( - self, - xdata: Incomplete, - ydata: Incomplete, - batch_size: Incomplete, - nepochs: Incomplete, - **kwargs: Incomplete, - ) -> None: ... diff --git a/stubs/lemminflect/kmodels/ModelLemmaClasses.pyi b/stubs/lemminflect/kmodels/ModelLemmaClasses.pyi deleted file mode 100644 index fcf6cfa..0000000 --- a/stubs/lemminflect/kmodels/ModelLemmaClasses.pyi +++ /dev/null @@ -1,21 +0,0 @@ -# Auto-generated by stubgen - -from _typeshed import Incomplete - -class ModelLemmaClasses: - rules: list[str] - rules_dict: dict[str, int] - def __init__(self, fn: Incomplete | None = None) -> None: ... - def getRuleIndex( - self, rule: tuple[Incomplete, Incomplete, Incomplete] | str - ) -> int: ... - @staticmethod - def computeSuffixRule( - infl: Incomplete, lemma: str - ) -> tuple[Incomplete, Incomplete, Incomplete]: ... - @classmethod - def saveFromRuleTuples( - cls: object, - fn: Incomplete, - rules: set[tuple[Incomplete, Incomplete, Incomplete]], - ) -> None: ... diff --git a/stubs/lemminflect/kmodels/ModelLemmaInData.pyi b/stubs/lemminflect/kmodels/ModelLemmaInData.pyi deleted file mode 100644 index 20c2890..0000000 --- a/stubs/lemminflect/kmodels/ModelLemmaInData.pyi +++ /dev/null @@ -1,23 +0,0 @@ -# Auto-generated by stubgen - -from typing import Any - -import numpy as np -from _typeshed import Incomplete - -from ..codecs.LemmaTCorpFileCodec import ( - LemmaTCorpFileCodec as LemmaTCorpFileCodec, -) -from ..slexicon.SKey import * # noqa: F403 - -class ModelLemmaInData: - WVEC_LEN: int - MAX_LETTER_IDX: int - entries: Incomplete - def __init__(self, fn: Incomplete) -> None: ... - @classmethod - def wordToVec( - cls: object, word: str, category: str - ) -> np.ndarray[Any, Any]: ... - @staticmethod - def getLetterClasses() -> list[str]: ... diff --git a/stubs/lemminflect/kmodels/__init__.pyi b/stubs/lemminflect/kmodels/__init__.pyi deleted file mode 100644 index f68e791..0000000 --- a/stubs/lemminflect/kmodels/__init__.pyi +++ /dev/null @@ -1 +0,0 @@ -# Auto-generated by stubgen diff --git a/stubs/lemminflect/slexicon/SKey.pyi b/stubs/lemminflect/slexicon/SKey.pyi deleted file mode 100644 index 1b568c4..0000000 --- a/stubs/lemminflect/slexicon/SKey.pyi +++ /dev/null @@ -1,48 +0,0 @@ -class SKey: - NOUN: str - ADJ: str - ADV: str - VERB: str - AUX: str - MODAL: str - PRON: str - DET: str - PREP: str - CONJ: str - COMPL: str - GENDER: str - STATIVE: str - PROPER: str - BROAD_NEG: str - INTERR: str - DEMONST: str - INTRAN: str - TRAN: str - DITRAN: str - LINK: str - CPLXTRAN: str - MOD_TYPE: str - POSITION: str - TYPE: str - IRREG: str - REG: str - REGD: str - GLREG: str - UNCOUNT: str - GCOUNT: str - METAREG: str - PLUR: str - SING: str - INV: str - INV_PERIPH: str - SINGULAR: str - PLURAL: str - INFINATIVE: str - THIRD_PRES: str - PAST: str - PAST_PART: str - PRES_PART: str - PRESENT: str - POSITIVE: str - COMPARATIVE: str - SUPERLATIVE: str diff --git a/stubs/lemminflect/slexicon/SPECIALISTEntry.pyi b/stubs/lemminflect/slexicon/SPECIALISTEntry.pyi deleted file mode 100644 index b984574..0000000 --- a/stubs/lemminflect/slexicon/SPECIALISTEntry.pyi +++ /dev/null @@ -1,44 +0,0 @@ -from typing import Any - -class StandardVariant: - vtype: Any - isgroup: Any - irreg: Any - def __init__( - self, - vtype: Any, - isgroup: bool = False, - irreg: Any | None = None, - ) -> None: ... - def __eq__(self, other: object) -> bool: ... - def __ne__(self, other: object) -> bool: ... - def __hash__(self) -> int: ... - -class AuxModVariant: - inflection: Any - form: Any - agreements: Any - negative: Any - def __init__( - self, - inflection: Any, - form: Any, - agreements: Any, - negative: Any, - ) -> None: ... - def __eq__(self, other: object) -> bool: ... - def __ne__(self, other: object) -> bool: ... - def __hash__(self) -> int: ... - -class SPECIALISTEntry: - EUI: Any - base: Any - category: Any - spelling_variant: Any - features: Any - acronym_of: Any - nominalization_of: Any - nominalization: Any - variants: Any - def __init__(self) -> None: ... - def getString(self) -> str: ... diff --git a/stubs/lemminflect/slexicon/SPECIALISTExtractor.pyi b/stubs/lemminflect/slexicon/SPECIALISTExtractor.pyi deleted file mode 100644 index 551cc98..0000000 --- a/stubs/lemminflect/slexicon/SPECIALISTExtractor.pyi +++ /dev/null @@ -1,11 +0,0 @@ -from typing import Any - -from .SPECIALISTEntry import AuxModVariant as AuxModVariant -from .SPECIALISTEntry import SPECIALISTEntry as SPECIALISTEntry -from .SPECIALISTEntry import StandardVariant as StandardVariant - -class SPECIALISTExtractor: - word_set: set[str] - lexicon: list[Any] - def __init__(self, word_set_fn: str | None = None) -> None: ... - def extract(self, lexicon_fn: str) -> None: ... diff --git a/stubs/lemminflect/slexicon/__init__.pyi b/stubs/lemminflect/slexicon/__init__.pyi deleted file mode 100644 index f68e791..0000000 --- a/stubs/lemminflect/slexicon/__init__.pyi +++ /dev/null @@ -1 +0,0 @@ -# Auto-generated by stubgen diff --git a/stubs/lemminflect/utils/CorpusUtils.pyi b/stubs/lemminflect/utils/CorpusUtils.pyi deleted file mode 100644 index 589813f..0000000 --- a/stubs/lemminflect/utils/CorpusUtils.pyi +++ /dev/null @@ -1,7 +0,0 @@ -import re - -def loadNLTKCorpus(corp_fn: str, max_chars: int = int(1e12)) -> list[str]: ... - -is_ascii_regex: re.Pattern[str] - -def isASCIIWord(word: str) -> bool: ... diff --git a/stubs/lemminflect/utils/DataContainer.pyi b/stubs/lemminflect/utils/DataContainer.pyi deleted file mode 100644 index 51452b0..0000000 --- a/stubs/lemminflect/utils/DataContainer.pyi +++ /dev/null @@ -1,9 +0,0 @@ -from gzip import GzipFile - -class DataContainer: - def __init__(self, obj: object | None = None) -> None: ... - def save(self, filename: str) -> None: ... - @staticmethod - def _open(filename: str, mode: str) -> GzipFile: ... - @classmethod - def load(cls: object, filename: str) -> DataContainer: ... diff --git a/stubs/lemminflect/utils/ProgressBar.pyi b/stubs/lemminflect/utils/ProgressBar.pyi deleted file mode 100644 index d175903..0000000 --- a/stubs/lemminflect/utils/ProgressBar.pyi +++ /dev/null @@ -1,6 +0,0 @@ -class ProgressBar: - end_val: int - bar_len: int - def __init__(self, end_val: int, bar_len: int = 20) -> None: ... - def update(self, val: int) -> None: ... - def clear(self) -> None: ... diff --git a/stubs/lemminflect/utils/Singleton.pyi b/stubs/lemminflect/utils/Singleton.pyi deleted file mode 100644 index c9ad834..0000000 --- a/stubs/lemminflect/utils/Singleton.pyi +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Any, Union - -from lemminflect.core.Inflections import Inflections -from lemminflect.core.Lemmatizer import Lemmatizer - -class _Singleton: - def __call__( - cls: object, *args: list[Any], **kwargs: list[Any] - ) -> Union[Inflections, Lemmatizer]: ... - -class Singleton(_Singleton): - pass diff --git a/stubs/lemminflect/utils/Unigrams.pyi b/stubs/lemminflect/utils/Unigrams.pyi deleted file mode 100644 index 62dbad7..0000000 --- a/stubs/lemminflect/utils/Unigrams.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from collections import Counter - -from ..slexicon.SKey import SKey - -class Unigrams: - counter: Counter[tuple[str, str]] - counter_word: Counter[str] - - def __init__(self, fn: str) -> None: ... - def getCount(self, word: str, tag: str | None = None) -> int: ... - def getCountForLemma(self, word: str, category: SKey) -> int: ... - def getCountForInflections( - self, word: str, category: SKey, infl_type: str - ) -> int: ... - def save(self, fn: str, min_count: int = 1) -> None: ... - @staticmethod - def saveCounter( - fn: str, counter: Counter[tuple[str, str]], min_count: int = 1 - ) -> None: ... - @staticmethod - def load(fn: str) -> Counter[tuple[str, str]]: ... - @staticmethod - def convertToWordKey( - counter: Counter[tuple[str, str]], - ) -> Counter[str]: ... diff --git a/stubs/lemminflect/utils/__init__.pyi b/stubs/lemminflect/utils/__init__.pyi deleted file mode 100644 index f68e791..0000000 --- a/stubs/lemminflect/utils/__init__.pyi +++ /dev/null @@ -1 +0,0 @@ -# Auto-generated by stubgen diff --git a/stubs/nltk/__init__.pyi b/stubs/nltk/__init__.pyi index 6eadc63..522d956 100644 --- a/stubs/nltk/__init__.pyi +++ b/stubs/nltk/__init__.pyi @@ -1,114 +1,12 @@ -from nltk import ( - ccg as ccg, -) -from nltk import ( - chunk as chunk, -) -from nltk import ( - classify as classify, -) -from nltk import ( - cluster as cluster, -) -from nltk import ( - collocations as collocations, -) -from nltk import ( - data as data, -) -from nltk import ( - featstruct as featstruct, -) -from nltk import ( - grammar as grammar, -) -from nltk import ( - help as help, -) -from nltk import ( - inference as inference, -) -from nltk import ( - lazyimport as lazyimport, -) -from nltk import ( - metrics as metrics, -) -from nltk import ( - misc as misc, -) -from nltk import ( - parse as parse, -) -from nltk import ( - probability as probability, -) -from nltk import ( - sem as sem, -) -from nltk import ( - stem as stem, -) -from nltk import ( - tag as tag, -) -from nltk import ( - tbl as tbl, -) -from nltk import ( - text as text, -) -from nltk import ( - tokenize as tokenize, -) -from nltk import ( - translate as translate, -) -from nltk import ( - tree as tree, -) -from nltk import ( - util as util, -) -from nltk import ( - wsd as wsd, -) -from nltk.chunk import * -from nltk.classify import * # type: ignore[assignment] -from nltk.collocations import * -from nltk.decorators import decorator as decorator -from nltk.decorators import memoize as memoize -from nltk.downloader import ( - download as download, -) -from nltk.downloader import ( - download_gui as download_gui, -) -from nltk.downloader import ( - download_shell as download_shell, -) -from nltk.featstruct import * -from nltk.grammar import * -from nltk.inference import * -from nltk.internals import config_java as config_java -from nltk.jsontags import * -from nltk.lazyimport import * -from nltk.metrics import * # type: ignore[assignment] -from nltk.parse import * -from nltk.probability import * # type: ignore[assignment] -from nltk.sem import * -from nltk.stem import * -from nltk.tag import * -from nltk.text import * -from nltk.tokenize import * -from nltk.translate import * -from nltk.tree import * -from nltk.util import * +from _typeshed import Incomplete -app: LazyModule -chat: LazyModule -corpus: LazyModule -draw: LazyModule -toolbox: LazyModule - -def demo() -> None: ... +def download( + info_or_id: str | None = None, + download_dir: str | None = None, + quiet: bool = False, + force: bool = False, + prefix: str = "[nltk_data] ", + halt_on_error: bool = True, + raise_on_error: bool = False, + print_error_to: Incomplete = ..., +) -> bool: ... diff --git a/stubs/nltk/allowlist b/stubs/nltk/allowlist index 07f4b10..1a71fef 100644 --- a/stubs/nltk/allowlist +++ b/stubs/nltk/allowlist @@ -1,95 +1,4 @@ -nltk.AlignedSent.alignment -nltk.AlignedSent.invert -nltk.Alignment.__getitem__ -nltk.Alignment.fromstring -nltk.Alignment.invert -nltk.Alignment.range -nltk.BigramAssocMeasures -nltk.BinaryMaxentFeatureEncoding.describe -nltk.BinaryMaxentFeatureEncoding.length -nltk.DictionaryConditionalProbDist -nltk.DictionaryConditionalProbDist.__missing__ -nltk.DictionaryProbDist -nltk.FeatStruct.__ge__ -nltk.FeatStruct.__gt__ -nltk.FeatStruct.__le__ -nltk.Feature.__ge__ -nltk.Feature.__gt__ -nltk.Feature.__le__ -nltk.IBMModel.MIN_PROB -nltk.IBMModel.maximize_fertility_probabilities -nltk.IBMModel.maximize_lexical_translation_probabilities -nltk.IBMModel.maximize_null_generation_probabilities -nltk.IBMModel.prob_of_alignments -nltk.IBMModel.prob_t_a_given_s -nltk.IBMModel.set_uniform_probabilities -nltk.IBMModel1.prob_alignment_point -nltk.IBMModel1.prob_all_alignments -nltk.IBMModel1.train -nltk.IBMModel2.maximize_alignment_probabilities -nltk.IBMModel2.prob_alignment_point -nltk.IBMModel2.prob_all_alignments -nltk.IBMModel2.train -nltk.IBMModel3.maximize_distortion_probabilities -nltk.IBMModel3.train -nltk.IBMModel4.maximize_distortion_probabilities -nltk.IBMModel4.train -nltk.IBMModel5.MIN_SCORE_FACTOR -nltk.IBMModel5.hillclimb -nltk.IBMModel5.maximize_vacancy_probabilities -nltk.IBMModel5.sample -nltk.IBMModel5.train -nltk.LazyModule -nltk.LineTokenizer -nltk.LineTokenizer.span_tokenize -nltk.LineTokenizer.tokenize -nltk.MaxentClassifier.ALGORITHMS -nltk.MaxentClassifier.__init__ -nltk.MaxentClassifier.classify -nltk.MaxentClassifier.explain -nltk.MaxentClassifier.labels -nltk.MaxentClassifier.most_informative_features -nltk.MaxentClassifier.prob_classify -nltk.MaxentClassifier.set_weights -nltk.MaxentClassifier.show_most_informative_features -nltk.MaxentClassifier.weights -nltk.MultiParentedTree -nltk.NgramAssocMeasures -nltk.Nonterminal.__ge__ -nltk.Nonterminal.__gt__ -nltk.Nonterminal.__le__ -nltk.ParentedTree -nltk.ParentedTree.copy -nltk.ProbabilisticNonprojectiveParser.best_incoming_arc -nltk.ProbabilisticNonprojectiveParser.original_best_arc -nltk.Production.__ge__ -nltk.Production.__gt__ -nltk.Production.__le__ -nltk.QuadgramAssocMeasures -nltk.RegexpChunkParser.rules -nltk.Senna -nltk.Senna.SUPPORTED_OPERATIONS -nltk.SennaChunkTagger -nltk.SennaNERTagger -nltk.SennaTagger -nltk.StackDecoder.distortion_factor -nltk.StackDecoder.expansion_score -nltk.StackDecoder.translate -nltk.StanfordSegmenter -nltk.SyllableTokenizer -nltk.TreebankWordTokenizer -nltk.TreebankWordTokenizer.CONTRACTIONS2 -nltk.TreebankWordTokenizer.CONTRACTIONS3 -nltk.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.TreebankWordTokenizer.DOUBLE_DASHES -nltk.TreebankWordTokenizer.ENDING_QUOTES -nltk.TreebankWordTokenizer.PARENS_BRACKETS -nltk.TreebankWordTokenizer.PUNCTUATION -nltk.TreebankWordTokenizer.STARTING_QUOTES -nltk.TrigramAssocMeasures -nltk.Variable.__ge__ -nltk.Variable.__gt__ -nltk.Variable.__le__ +nltk.PRETRAINED_TAGGERS nltk.__author_email__ nltk.__classifiers__ nltk.__keywords__ @@ -98,147 +7,50 @@ nltk.__longdescr__ nltk.__maintainer__ nltk.__maintainer_email__ nltk.__url__ -nltk.accuracy -nltk.app.chartparser_app.Nonterminal.__ge__ -nltk.app.chartparser_app.Nonterminal.__gt__ -nltk.app.chartparser_app.Nonterminal.__le__ -nltk.app.chartparser_app.TreeEdge.move_dot_forward -nltk.app.chartparser_app.__all__ -nltk.app.chunkparser_app.RegexpChunkParser.rules -nltk.app.chunkparser_app.RegexpChunkRule.descr -nltk.app.chunkparser_app.__all__ -nltk.app.chunkparser_app.conll2000 -nltk.app.chunkparser_app.treebank_chunk -nltk.app.collocations_app.__all__ -nltk.app.collocations_app.alpino -nltk.app.collocations_app.brown -nltk.app.collocations_app.cess_cat -nltk.app.collocations_app.cess_esp -nltk.app.collocations_app.floresta -nltk.app.collocations_app.indian -nltk.app.collocations_app.mac_morpho -nltk.app.collocations_app.machado -nltk.app.collocations_app.nps_chat -nltk.app.collocations_app.sinica_treebank -nltk.app.collocations_app.treebank -nltk.app.concordance_app.__all__ -nltk.app.concordance_app.alpino -nltk.app.concordance_app.brown -nltk.app.concordance_app.cess_cat -nltk.app.concordance_app.cess_esp -nltk.app.concordance_app.floresta -nltk.app.concordance_app.indian -nltk.app.concordance_app.mac_morpho -nltk.app.concordance_app.nps_chat -nltk.app.concordance_app.sinica_treebank -nltk.app.concordance_app.treebank -nltk.app.nemo_app.__all__ -nltk.app.pylab -nltk.app.rdparser_app.__all__ -nltk.app.srparser_app.__all__ +nltk.app +nltk.app.chartparser_app +nltk.app.chunkparser_app +nltk.app.collocations_app +nltk.app.concordance_app +nltk.app.nemo_app +nltk.app.rdparser_app +nltk.app.srparser_app nltk.app.wordfreq_app +nltk.app.wordnet_app nltk.book -nltk.ccg.api.AbstractCCGCategory.__ge__ -nltk.ccg.api.AbstractCCGCategory.__gt__ -nltk.ccg.api.AbstractCCGCategory.__le__ -nltk.ccg.api.AbstractCCGCategory.__str__ -nltk.ccg.api.Direction.__ge__ -nltk.ccg.api.Direction.__gt__ -nltk.ccg.api.Direction.__le__ -nltk.ccg.chart.AbstractVariableExpression.__ge__ -nltk.ccg.chart.AbstractVariableExpression.__le__ -nltk.ccg.chart.EdgeI.__ge__ -nltk.ccg.chart.EdgeI.__gt__ -nltk.ccg.chart.EdgeI.__le__ -nltk.ccg.chart.Variable.__ge__ -nltk.ccg.chart.Variable.__gt__ -nltk.ccg.chart.Variable.__le__ -nltk.ccg.chart.demo-redefinition -nltk.ccg.chart.innermostFunction -nltk.ccg.combinator.innermostFunction -nltk.ccg.lexicon.Direction.__ge__ -nltk.ccg.lexicon.Direction.__gt__ -nltk.ccg.lexicon.Direction.__le__ -nltk.ccg.lexicon.parseApplication -nltk.ccg.logic.AbstractVariableExpression.__ge__ -nltk.ccg.logic.AbstractVariableExpression.__le__ -nltk.ccg.logic.Variable.__ge__ -nltk.ccg.logic.Variable.__gt__ -nltk.ccg.logic.Variable.__le__ -nltk.chunk.RegexpChunkParser.rules -nltk.chunk.named_entity.MaxentClassifier.ALGORITHMS -nltk.chunk.named_entity.MaxentClassifier.__init__ -nltk.chunk.named_entity.MaxentClassifier.classify -nltk.chunk.named_entity.MaxentClassifier.explain -nltk.chunk.named_entity.MaxentClassifier.labels -nltk.chunk.named_entity.MaxentClassifier.most_informative_features -nltk.chunk.named_entity.MaxentClassifier.prob_classify -nltk.chunk.named_entity.MaxentClassifier.set_weights -nltk.chunk.named_entity.MaxentClassifier.show_most_informative_features -nltk.chunk.named_entity.MaxentClassifier.weights -nltk.chunk.regexp.ChunkString.CHUNK_TAG -nltk.chunk.regexp.ChunkString.CHUNK_TAG_CHAR -nltk.chunk.regexp.ChunkString.IN_CHUNK_PATTERN -nltk.chunk.regexp.ChunkString.IN_STRIP_PATTERN -nltk.chunk.regexp.RegexpChunkParser.rules -nltk.chunk.regexp.RegexpChunkRule.descr -nltk.classify.BinaryMaxentFeatureEncoding.describe -nltk.classify.BinaryMaxentFeatureEncoding.length -nltk.classify.MaxentClassifier.ALGORITHMS -nltk.classify.MaxentClassifier.__init__ -nltk.classify.MaxentClassifier.classify -nltk.classify.MaxentClassifier.explain -nltk.classify.MaxentClassifier.labels -nltk.classify.MaxentClassifier.most_informative_features -nltk.classify.MaxentClassifier.prob_classify -nltk.classify.MaxentClassifier.set_weights -nltk.classify.MaxentClassifier.show_most_informative_features -nltk.classify.MaxentClassifier.weights -nltk.classify.Senna -nltk.classify.Senna.SUPPORTED_OPERATIONS -nltk.classify.maxent.BinaryMaxentFeatureEncoding.describe -nltk.classify.maxent.BinaryMaxentFeatureEncoding.length -nltk.classify.maxent.ConditionalExponentialClassifier -nltk.classify.maxent.FunctionBackedMaxentFeatureEncoding -nltk.classify.maxent.GISEncoding -nltk.classify.maxent.MaxentClassifier.ALGORITHMS -nltk.classify.maxent.MaxentClassifier.__init__ -nltk.classify.maxent.MaxentClassifier.classify -nltk.classify.maxent.MaxentClassifier.explain -nltk.classify.maxent.MaxentClassifier.labels -nltk.classify.maxent.MaxentClassifier.most_informative_features -nltk.classify.maxent.MaxentClassifier.prob_classify -nltk.classify.maxent.MaxentClassifier.set_weights -nltk.classify.maxent.MaxentClassifier.show_most_informative_features -nltk.classify.maxent.MaxentClassifier.weights -nltk.classify.maxent.MaxentFeatureEncodingI -nltk.classify.maxent.TadmEventMaxentFeatureEncoding.describe -nltk.classify.maxent.TadmEventMaxentFeatureEncoding.length -nltk.classify.maxent.TypedMaxentFeatureEncoding -nltk.classify.maxent.__docformat__ -nltk.classify.maxent.calculate_deltas -nltk.classify.maxent.calculate_empirical_fcount -nltk.classify.maxent.calculate_estimated_fcount -nltk.classify.maxent.calculate_nfmap -nltk.classify.maxent.demo -nltk.classify.maxent.load_maxent_params -nltk.classify.maxent.maxent_pos_tagger -nltk.classify.maxent.save_maxent_params -nltk.classify.maxent.train_maxent_classifier_with_gis -nltk.classify.maxent.train_maxent_classifier_with_iis -nltk.classify.megam.parse_megam_weights -nltk.classify.naivebayes.DictionaryProbDist -nltk.classify.positivenaivebayes.DictionaryProbDist -nltk.classify.rte_classify.RTEFeatureExtractor -nltk.classify.rte_classify.rte_features -nltk.classify.rte_classify.rte_featurize -nltk.classify.senna.Senna -nltk.classify.senna.Senna.SUPPORTED_OPERATIONS +nltk.ccg +nltk.ccg.api +nltk.ccg.chart +nltk.ccg.combinator +nltk.ccg.lexicon +nltk.ccg.logic +nltk.chat +nltk.chat.eliza +nltk.chat.iesha +nltk.chat.rude +nltk.chat.suntsu +nltk.chat.util +nltk.chat.zen +nltk.chunk +nltk.chunk.api +nltk.chunk.named_entity +nltk.chunk.regexp +nltk.chunk.util +nltk.classify +nltk.classify.api +nltk.classify.decisiontree +nltk.classify.maxent +nltk.classify.megam +nltk.classify.naivebayes +nltk.classify.positivenaivebayes +nltk.classify.rte_classify +nltk.classify.scikitlearn +nltk.classify.senna nltk.classify.svm -nltk.classify.tadm.encoding_demo -nltk.classify.tadm.names_demo -nltk.classify.tadm.parse_tadm_weights -nltk.classify.weka.DictionaryProbDist +nltk.classify.tadm +nltk.classify.textcat +nltk.classify.util +nltk.classify.weka nltk.cli nltk.cluster nltk.cluster.api @@ -247,36 +59,11 @@ nltk.cluster.gaac nltk.cluster.kmeans nltk.cluster.util nltk.collections -nltk.corpus.CategorizedBracketParseCorpusReader.parsed_paras -nltk.corpus.CategorizedBracketParseCorpusReader.parsed_sents -nltk.corpus.CategorizedBracketParseCorpusReader.parsed_words -nltk.corpus.CategorizedBracketParseCorpusReader.tagged_paras -nltk.corpus.CategorizedBracketParseCorpusReader.tagged_sents -nltk.corpus.CategorizedCorpusReader.paras -nltk.corpus.CategorizedCorpusReader.raw -nltk.corpus.CategorizedCorpusReader.sents -nltk.corpus.CategorizedSentencesCorpusReader.CorpusView -nltk.corpus.ComparativeSentencesCorpusReader.CorpusView -nltk.corpus.CorpusReader.citation -nltk.corpus.CorpusReader.license -nltk.corpus.CorpusReader.raw -nltk.corpus.CorpusReader.readme -nltk.corpus.CorpusReader.root -nltk.corpus.DependencyCorpusReader.__init__ -nltk.corpus.DependencyCorpusReader.tagged_sents -nltk.corpus.DependencyCorpusReader.tagged_words -nltk.corpus.DependencyCorpusReader.words -nltk.corpus.NombankCorpusReader.instances -nltk.corpus.NombankCorpusReader.lines -nltk.corpus.NombankCorpusReader.roleset -nltk.corpus.NombankCorpusReader.rolesets -nltk.corpus.OpinionLexiconCorpusReader.CorpusView -nltk.corpus.PlaintextCorpusReader.CorpusView -nltk.corpus.ProsConsCorpusReader.CorpusView -nltk.corpus.ReviewsCorpusReader.CorpusView -nltk.corpus.SyntaxCorpusReader.parsed_sents -nltk.corpus.SyntaxCorpusReader.sents -nltk.corpus.TwitterCorpusReader.CorpusView +nltk.collocations +nltk.compat +nltk.corpus +nltk.corpus.Lemma +nltk.corpus.Synset nltk.corpus.WordNetCorpusReader.ADJ nltk.corpus.WordNetCorpusReader.ADJ_SAT nltk.corpus.WordNetCorpusReader.ADV @@ -331,89 +118,9 @@ nltk.corpus.conll2000 nltk.corpus.conll2002 nltk.corpus.conll2007 nltk.corpus.crubadan +nltk.corpus.demo nltk.corpus.dependency_treebank -nltk.corpus.europarl_raw.CategorizedBracketParseCorpusReader.parsed_paras -nltk.corpus.europarl_raw.CategorizedBracketParseCorpusReader.parsed_sents -nltk.corpus.europarl_raw.CategorizedBracketParseCorpusReader.parsed_words -nltk.corpus.europarl_raw.CategorizedBracketParseCorpusReader.tagged_paras -nltk.corpus.europarl_raw.CategorizedBracketParseCorpusReader.tagged_sents -nltk.corpus.europarl_raw.CategorizedCorpusReader.paras -nltk.corpus.europarl_raw.CategorizedCorpusReader.raw -nltk.corpus.europarl_raw.CategorizedCorpusReader.sents -nltk.corpus.europarl_raw.CategorizedSentencesCorpusReader.CorpusView -nltk.corpus.europarl_raw.ComparativeSentencesCorpusReader.CorpusView -nltk.corpus.europarl_raw.CorpusReader.citation -nltk.corpus.europarl_raw.CorpusReader.license -nltk.corpus.europarl_raw.CorpusReader.raw -nltk.corpus.europarl_raw.CorpusReader.readme -nltk.corpus.europarl_raw.CorpusReader.root -nltk.corpus.europarl_raw.DependencyCorpusReader.__init__ -nltk.corpus.europarl_raw.DependencyCorpusReader.tagged_sents -nltk.corpus.europarl_raw.DependencyCorpusReader.tagged_words -nltk.corpus.europarl_raw.DependencyCorpusReader.words -nltk.corpus.europarl_raw.NombankCorpusReader.instances -nltk.corpus.europarl_raw.NombankCorpusReader.lines -nltk.corpus.europarl_raw.NombankCorpusReader.roleset -nltk.corpus.europarl_raw.NombankCorpusReader.rolesets -nltk.corpus.europarl_raw.OpinionLexiconCorpusReader.CorpusView -nltk.corpus.europarl_raw.PlaintextCorpusReader.CorpusView -nltk.corpus.europarl_raw.ProsConsCorpusReader.CorpusView -nltk.corpus.europarl_raw.ReviewsCorpusReader.CorpusView -nltk.corpus.europarl_raw.SyntaxCorpusReader.parsed_sents -nltk.corpus.europarl_raw.SyntaxCorpusReader.sents -nltk.corpus.europarl_raw.TwitterCorpusReader.CorpusView -nltk.corpus.europarl_raw.WordNetCorpusReader.ADJ -nltk.corpus.europarl_raw.WordNetCorpusReader.ADJ_SAT -nltk.corpus.europarl_raw.WordNetCorpusReader.ADV -nltk.corpus.europarl_raw.WordNetCorpusReader.MORPHOLOGICAL_SUBSTITUTIONS -nltk.corpus.europarl_raw.WordNetCorpusReader.NOUN -nltk.corpus.europarl_raw.WordNetCorpusReader.VERB -nltk.corpus.europarl_raw.WordNetCorpusReader.add_exomw -nltk.corpus.europarl_raw.WordNetCorpusReader.add_omw -nltk.corpus.europarl_raw.WordNetCorpusReader.add_provs -nltk.corpus.europarl_raw.WordNetCorpusReader.all_eng_synsets -nltk.corpus.europarl_raw.WordNetCorpusReader.all_lemma_names -nltk.corpus.europarl_raw.WordNetCorpusReader.all_omw_synsets -nltk.corpus.europarl_raw.WordNetCorpusReader.all_synsets -nltk.corpus.europarl_raw.WordNetCorpusReader.citation -nltk.corpus.europarl_raw.WordNetCorpusReader.custom_lemmas -nltk.corpus.europarl_raw.WordNetCorpusReader.digraph -nltk.corpus.europarl_raw.WordNetCorpusReader.disable_custom_lemmas -nltk.corpus.europarl_raw.WordNetCorpusReader.doc -nltk.corpus.europarl_raw.WordNetCorpusReader.ic -nltk.corpus.europarl_raw.WordNetCorpusReader.jcn_similarity -nltk.corpus.europarl_raw.WordNetCorpusReader.langs -nltk.corpus.europarl_raw.WordNetCorpusReader.lch_similarity -nltk.corpus.europarl_raw.WordNetCorpusReader.lemma -nltk.corpus.europarl_raw.WordNetCorpusReader.lemma_count -nltk.corpus.europarl_raw.WordNetCorpusReader.lemma_from_key -nltk.corpus.europarl_raw.WordNetCorpusReader.lemmas -nltk.corpus.europarl_raw.WordNetCorpusReader.license -nltk.corpus.europarl_raw.WordNetCorpusReader.lin_similarity -nltk.corpus.europarl_raw.WordNetCorpusReader.merged_synsets -nltk.corpus.europarl_raw.WordNetCorpusReader.morphy -nltk.corpus.europarl_raw.WordNetCorpusReader.of2ss -nltk.corpus.europarl_raw.WordNetCorpusReader.path_similarity -nltk.corpus.europarl_raw.WordNetCorpusReader.readme -nltk.corpus.europarl_raw.WordNetCorpusReader.res_similarity -nltk.corpus.europarl_raw.WordNetCorpusReader.split_synsets -nltk.corpus.europarl_raw.WordNetCorpusReader.ss2of -nltk.corpus.europarl_raw.WordNetCorpusReader.synonyms -nltk.corpus.europarl_raw.WordNetCorpusReader.synset -nltk.corpus.europarl_raw.WordNetCorpusReader.synset_from_sense_key -nltk.corpus.europarl_raw.WordNetCorpusReader.words -nltk.corpus.europarl_raw.WordNetCorpusReader.wup_similarity -nltk.corpus.europarl_raw.danish -nltk.corpus.europarl_raw.dutch -nltk.corpus.europarl_raw.english -nltk.corpus.europarl_raw.finnish -nltk.corpus.europarl_raw.french -nltk.corpus.europarl_raw.german -nltk.corpus.europarl_raw.greek -nltk.corpus.europarl_raw.italian -nltk.corpus.europarl_raw.portuguese -nltk.corpus.europarl_raw.spanish -nltk.corpus.europarl_raw.swedish +nltk.corpus.europarl_raw nltk.corpus.extended_omw nltk.corpus.floresta nltk.corpus.framenet @@ -447,34 +154,32 @@ nltk.corpus.propbank_ptb nltk.corpus.pros_cons nltk.corpus.ptb nltk.corpus.qc -nltk.corpus.reader.AlignedSent.alignment -nltk.corpus.reader.AlignedSent.invert -nltk.corpus.reader.Alignment.__getitem__ -nltk.corpus.reader.Alignment.fromstring -nltk.corpus.reader.Alignment.invert -nltk.corpus.reader.Alignment.range -nltk.corpus.reader.CategorizedBracketParseCorpusReader.parsed_paras -nltk.corpus.reader.CategorizedBracketParseCorpusReader.parsed_sents -nltk.corpus.reader.CategorizedBracketParseCorpusReader.parsed_words -nltk.corpus.reader.CategorizedBracketParseCorpusReader.tagged_paras -nltk.corpus.reader.CategorizedBracketParseCorpusReader.tagged_sents -nltk.corpus.reader.CategorizedCorpusReader.paras -nltk.corpus.reader.CategorizedCorpusReader.raw -nltk.corpus.reader.CategorizedCorpusReader.sents -nltk.corpus.reader.CategorizedSentencesCorpusReader.CorpusView -nltk.corpus.reader.ComparativeSentencesCorpusReader.CorpusView -nltk.corpus.reader.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.ConcatenatedCorpusView.close -nltk.corpus.reader.CorpusReader.citation -nltk.corpus.reader.CorpusReader.license -nltk.corpus.reader.CorpusReader.raw -nltk.corpus.reader.CorpusReader.readme -nltk.corpus.reader.CorpusReader.root -nltk.corpus.reader.DependencyCorpusReader.__init__ -nltk.corpus.reader.DependencyCorpusReader.tagged_sents -nltk.corpus.reader.DependencyCorpusReader.tagged_words -nltk.corpus.reader.DependencyCorpusReader.words +nltk.corpus.reader.AlignedCorpusReader +nltk.corpus.reader.AlpinoCorpusReader +nltk.corpus.reader.BCP47CorpusReader +nltk.corpus.reader.BNCCorpusReader +nltk.corpus.reader.BracketParseCorpusReader +nltk.corpus.reader.CHILDESCorpusReader +nltk.corpus.reader.CMUDictCorpusReader +nltk.corpus.reader.CategorizedBracketParseCorpusReader +nltk.corpus.reader.CategorizedCorpusReader +nltk.corpus.reader.CategorizedPlaintextCorpusReader +nltk.corpus.reader.CategorizedSentencesCorpusReader +nltk.corpus.reader.CategorizedTaggedCorpusReader +nltk.corpus.reader.ChasenCorpusReader +nltk.corpus.reader.ChunkedCorpusReader +nltk.corpus.reader.ComparativeSentencesCorpusReader +nltk.corpus.reader.ConllChunkCorpusReader +nltk.corpus.reader.ConllCorpusReader +nltk.corpus.reader.CorpusReader +nltk.corpus.reader.CrubadanCorpusReader +nltk.corpus.reader.DependencyCorpusReader +nltk.corpus.reader.EuroparlCorpusReader +nltk.corpus.reader.FramenetCorpusReader +nltk.corpus.reader.IEERCorpusReader nltk.corpus.reader.IPIPANCorpusReader +nltk.corpus.reader.IndianCorpusReader +nltk.corpus.reader.KNBCorpusReader nltk.corpus.reader.Lemma.antonyms nltk.corpus.reader.Lemma.count nltk.corpus.reader.Lemma.derivationally_related_forms @@ -485,33 +190,33 @@ nltk.corpus.reader.Lemma.lang nltk.corpus.reader.Lemma.pertainyms nltk.corpus.reader.Lemma.synset nltk.corpus.reader.Lemma.syntactic_marker -nltk.corpus.reader.LineTokenizer -nltk.corpus.reader.LineTokenizer.span_tokenize -nltk.corpus.reader.LineTokenizer.tokenize +nltk.corpus.reader.LinThesaurusCorpusReader +nltk.corpus.reader.MTECorpusReader +nltk.corpus.reader.MWAPPDBCorpusReader +nltk.corpus.reader.MacMorphoCorpusReader nltk.corpus.reader.NKJPCorpusReader -nltk.corpus.reader.NombankCorpusReader.instances -nltk.corpus.reader.NombankCorpusReader.lines -nltk.corpus.reader.NombankCorpusReader.roleset -nltk.corpus.reader.NombankCorpusReader.rolesets -nltk.corpus.reader.OpinionLexiconCorpusReader.CorpusView -nltk.corpus.reader.PlaintextCorpusReader.CorpusView -nltk.corpus.reader.PropbankInstance.tree -nltk.corpus.reader.PropbankTreePointer.__ge__ -nltk.corpus.reader.PropbankTreePointer.__gt__ -nltk.corpus.reader.PropbankTreePointer.__le__ -nltk.corpus.reader.ProsConsCorpusReader.CorpusView -nltk.corpus.reader.ReviewsCorpusReader.CorpusView -nltk.corpus.reader.StanfordSegmenter -nltk.corpus.reader.StreamBackedCorpusView.__add__ -nltk.corpus.reader.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.StreamBackedCorpusView.__len__ -nltk.corpus.reader.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.StreamBackedCorpusView.fileid -nltk.corpus.reader.StreamBackedCorpusView.read_block -nltk.corpus.reader.SyllableTokenizer +nltk.corpus.reader.NPSChatCorpusReader +nltk.corpus.reader.NombankCorpusReader +nltk.corpus.reader.NonbreakingPrefixesCorpusReader +nltk.corpus.reader.OpinionLexiconCorpusReader +nltk.corpus.reader.PPAttachmentCorpusReader +nltk.corpus.reader.PanLexLiteCorpusReader +nltk.corpus.reader.PanlexSwadeshCorpusReader +nltk.corpus.reader.Pl196xCorpusReader +nltk.corpus.reader.PlaintextCorpusReader +nltk.corpus.reader.PortugueseCategorizedPlaintextCorpusReader +nltk.corpus.reader.PropbankCorpusReader +nltk.corpus.reader.ProsConsCorpusReader +nltk.corpus.reader.RTECorpusReader +nltk.corpus.reader.ReviewsCorpusReader +nltk.corpus.reader.SemcorCorpusReader +nltk.corpus.reader.SensevalCorpusReader +nltk.corpus.reader.SentiSynset +nltk.corpus.reader.SentiWordNetCorpusReader +nltk.corpus.reader.SinicaTreebankCorpusReader +nltk.corpus.reader.StringCategoryCorpusReader +nltk.corpus.reader.SwadeshCorpusReader +nltk.corpus.reader.SwitchboardCorpusReader nltk.corpus.reader.Synset.acyclic_tree nltk.corpus.reader.Synset.closure nltk.corpus.reader.Synset.common_hypernyms @@ -538,18 +243,17 @@ nltk.corpus.reader.Synset.root_hypernyms nltk.corpus.reader.Synset.shortest_path_distance nltk.corpus.reader.Synset.tree nltk.corpus.reader.Synset.wup_similarity -nltk.corpus.reader.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.SyntaxCorpusReader.sents -nltk.corpus.reader.TreebankWordTokenizer -nltk.corpus.reader.TreebankWordTokenizer.CONTRACTIONS2 -nltk.corpus.reader.TreebankWordTokenizer.CONTRACTIONS3 -nltk.corpus.reader.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.corpus.reader.TreebankWordTokenizer.DOUBLE_DASHES -nltk.corpus.reader.TreebankWordTokenizer.ENDING_QUOTES -nltk.corpus.reader.TreebankWordTokenizer.PARENS_BRACKETS -nltk.corpus.reader.TreebankWordTokenizer.PUNCTUATION -nltk.corpus.reader.TreebankWordTokenizer.STARTING_QUOTES -nltk.corpus.reader.TwitterCorpusReader.CorpusView +nltk.corpus.reader.SyntaxCorpusReader +nltk.corpus.reader.TEICorpusView +nltk.corpus.reader.TaggedCorpusReader +nltk.corpus.reader.TimitCorpusReader +nltk.corpus.reader.TimitTaggedCorpusReader +nltk.corpus.reader.ToolboxCorpusReader +nltk.corpus.reader.TwitterCorpusReader +nltk.corpus.reader.UdhrCorpusReader +nltk.corpus.reader.UnicharsCorpusReader +nltk.corpus.reader.VerbnetCorpusReader +nltk.corpus.reader.WordListCorpusReader nltk.corpus.reader.WordNetCorpusReader.ADJ nltk.corpus.reader.WordNetCorpusReader.ADJ_SAT nltk.corpus.reader.WordNetCorpusReader.ADV @@ -591,785 +295,65 @@ nltk.corpus.reader.WordNetCorpusReader.synset nltk.corpus.reader.WordNetCorpusReader.synset_from_sense_key nltk.corpus.reader.WordNetCorpusReader.words nltk.corpus.reader.WordNetCorpusReader.wup_similarity +nltk.corpus.reader.WordNetICCorpusReader +nltk.corpus.reader.XMLCorpusReader +nltk.corpus.reader.YCOECorpusReader nltk.corpus.reader.__all__ -nltk.corpus.reader.aligned.AlignedSent.alignment -nltk.corpus.reader.aligned.AlignedSent.invert -nltk.corpus.reader.aligned.Alignment.__getitem__ -nltk.corpus.reader.aligned.Alignment.fromstring -nltk.corpus.reader.aligned.Alignment.invert -nltk.corpus.reader.aligned.Alignment.range -nltk.corpus.reader.aligned.CorpusReader.citation -nltk.corpus.reader.aligned.CorpusReader.license -nltk.corpus.reader.aligned.CorpusReader.raw -nltk.corpus.reader.aligned.CorpusReader.readme -nltk.corpus.reader.aligned.CorpusReader.root -nltk.corpus.reader.aligned.StreamBackedCorpusView.__add__ -nltk.corpus.reader.aligned.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.aligned.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.aligned.StreamBackedCorpusView.__len__ -nltk.corpus.reader.aligned.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.aligned.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.aligned.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.aligned.StreamBackedCorpusView.fileid -nltk.corpus.reader.aligned.StreamBackedCorpusView.read_block -nltk.corpus.reader.api.CategorizedCorpusReader.paras -nltk.corpus.reader.api.CategorizedCorpusReader.raw -nltk.corpus.reader.api.CategorizedCorpusReader.sents +nltk.corpus.reader.aligned +nltk.corpus.reader.api.CategorizedCorpusReader nltk.corpus.reader.api.CorpusReader.citation nltk.corpus.reader.api.CorpusReader.license nltk.corpus.reader.api.CorpusReader.raw nltk.corpus.reader.api.CorpusReader.readme nltk.corpus.reader.api.CorpusReader.root -nltk.corpus.reader.api.StreamBackedCorpusView.__add__ -nltk.corpus.reader.api.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.api.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.api.StreamBackedCorpusView.__len__ -nltk.corpus.reader.api.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.api.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.api.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.api.StreamBackedCorpusView.fileid -nltk.corpus.reader.api.StreamBackedCorpusView.read_block -nltk.corpus.reader.api.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.api.SyntaxCorpusReader.sents -nltk.corpus.reader.bcp47.CorpusReader.citation -nltk.corpus.reader.bcp47.CorpusReader.license -nltk.corpus.reader.bcp47.CorpusReader.raw -nltk.corpus.reader.bcp47.CorpusReader.readme -nltk.corpus.reader.bcp47.CorpusReader.root -nltk.corpus.reader.bracket_parse.CategorizedBracketParseCorpusReader.parsed_paras -nltk.corpus.reader.bracket_parse.CategorizedBracketParseCorpusReader.parsed_sents -nltk.corpus.reader.bracket_parse.CategorizedBracketParseCorpusReader.parsed_words -nltk.corpus.reader.bracket_parse.CategorizedBracketParseCorpusReader.tagged_paras -nltk.corpus.reader.bracket_parse.CategorizedBracketParseCorpusReader.tagged_sents -nltk.corpus.reader.bracket_parse.EMPTY_BRACKETS -nltk.corpus.reader.bracket_parse.SORTTAGWRD -nltk.corpus.reader.bracket_parse.TAGWORD -nltk.corpus.reader.bracket_parse.WORD -nltk.corpus.reader.categorized_sents.CategorizedCorpusReader.paras -nltk.corpus.reader.categorized_sents.CategorizedCorpusReader.raw -nltk.corpus.reader.categorized_sents.CategorizedCorpusReader.sents -nltk.corpus.reader.categorized_sents.CategorizedSentencesCorpusReader.CorpusView -nltk.corpus.reader.categorized_sents.CorpusReader.citation -nltk.corpus.reader.categorized_sents.CorpusReader.license -nltk.corpus.reader.categorized_sents.CorpusReader.raw -nltk.corpus.reader.categorized_sents.CorpusReader.readme -nltk.corpus.reader.categorized_sents.CorpusReader.root -nltk.corpus.reader.categorized_sents.LineTokenizer -nltk.corpus.reader.categorized_sents.LineTokenizer.span_tokenize -nltk.corpus.reader.categorized_sents.LineTokenizer.tokenize -nltk.corpus.reader.categorized_sents.StanfordSegmenter -nltk.corpus.reader.categorized_sents.StreamBackedCorpusView.__add__ -nltk.corpus.reader.categorized_sents.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.categorized_sents.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.categorized_sents.StreamBackedCorpusView.__len__ -nltk.corpus.reader.categorized_sents.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.categorized_sents.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.categorized_sents.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.categorized_sents.StreamBackedCorpusView.fileid -nltk.corpus.reader.categorized_sents.StreamBackedCorpusView.read_block -nltk.corpus.reader.categorized_sents.SyllableTokenizer -nltk.corpus.reader.categorized_sents.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.categorized_sents.SyntaxCorpusReader.sents -nltk.corpus.reader.categorized_sents.TreebankWordTokenizer -nltk.corpus.reader.categorized_sents.TreebankWordTokenizer.CONTRACTIONS2 -nltk.corpus.reader.categorized_sents.TreebankWordTokenizer.CONTRACTIONS3 -nltk.corpus.reader.categorized_sents.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.corpus.reader.categorized_sents.TreebankWordTokenizer.DOUBLE_DASHES -nltk.corpus.reader.categorized_sents.TreebankWordTokenizer.ENDING_QUOTES -nltk.corpus.reader.categorized_sents.TreebankWordTokenizer.PARENS_BRACKETS -nltk.corpus.reader.categorized_sents.TreebankWordTokenizer.PUNCTUATION -nltk.corpus.reader.categorized_sents.TreebankWordTokenizer.STARTING_QUOTES -nltk.corpus.reader.chasen.CategorizedCorpusReader.paras -nltk.corpus.reader.chasen.CategorizedCorpusReader.raw -nltk.corpus.reader.chasen.CategorizedCorpusReader.sents -nltk.corpus.reader.chasen.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.chasen.ConcatenatedCorpusView.close -nltk.corpus.reader.chasen.CorpusReader.citation -nltk.corpus.reader.chasen.CorpusReader.license -nltk.corpus.reader.chasen.CorpusReader.raw -nltk.corpus.reader.chasen.CorpusReader.readme -nltk.corpus.reader.chasen.CorpusReader.root -nltk.corpus.reader.chasen.StreamBackedCorpusView.__add__ -nltk.corpus.reader.chasen.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.chasen.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.chasen.StreamBackedCorpusView.__len__ -nltk.corpus.reader.chasen.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.chasen.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.chasen.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.chasen.StreamBackedCorpusView.fileid -nltk.corpus.reader.chasen.StreamBackedCorpusView.read_block -nltk.corpus.reader.chasen.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.chasen.SyntaxCorpusReader.sents -nltk.corpus.reader.chunked.CategorizedCorpusReader.paras -nltk.corpus.reader.chunked.CategorizedCorpusReader.raw -nltk.corpus.reader.chunked.CategorizedCorpusReader.sents -nltk.corpus.reader.chunked.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.chunked.ConcatenatedCorpusView.close -nltk.corpus.reader.chunked.CorpusReader.citation -nltk.corpus.reader.chunked.CorpusReader.license -nltk.corpus.reader.chunked.CorpusReader.raw -nltk.corpus.reader.chunked.CorpusReader.readme -nltk.corpus.reader.chunked.CorpusReader.root -nltk.corpus.reader.chunked.LineTokenizer -nltk.corpus.reader.chunked.LineTokenizer.span_tokenize -nltk.corpus.reader.chunked.LineTokenizer.tokenize -nltk.corpus.reader.chunked.StanfordSegmenter -nltk.corpus.reader.chunked.StreamBackedCorpusView.__add__ -nltk.corpus.reader.chunked.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.chunked.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.chunked.StreamBackedCorpusView.__len__ -nltk.corpus.reader.chunked.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.chunked.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.chunked.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.chunked.StreamBackedCorpusView.fileid -nltk.corpus.reader.chunked.StreamBackedCorpusView.read_block -nltk.corpus.reader.chunked.SyllableTokenizer -nltk.corpus.reader.chunked.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.chunked.SyntaxCorpusReader.sents -nltk.corpus.reader.chunked.TreebankWordTokenizer -nltk.corpus.reader.chunked.TreebankWordTokenizer.CONTRACTIONS2 -nltk.corpus.reader.chunked.TreebankWordTokenizer.CONTRACTIONS3 -nltk.corpus.reader.chunked.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.corpus.reader.chunked.TreebankWordTokenizer.DOUBLE_DASHES -nltk.corpus.reader.chunked.TreebankWordTokenizer.ENDING_QUOTES -nltk.corpus.reader.chunked.TreebankWordTokenizer.PARENS_BRACKETS -nltk.corpus.reader.chunked.TreebankWordTokenizer.PUNCTUATION -nltk.corpus.reader.chunked.TreebankWordTokenizer.STARTING_QUOTES -nltk.corpus.reader.cmudict.CategorizedCorpusReader.paras -nltk.corpus.reader.cmudict.CategorizedCorpusReader.raw -nltk.corpus.reader.cmudict.CategorizedCorpusReader.sents -nltk.corpus.reader.cmudict.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.cmudict.ConcatenatedCorpusView.close -nltk.corpus.reader.cmudict.CorpusReader.citation -nltk.corpus.reader.cmudict.CorpusReader.license -nltk.corpus.reader.cmudict.CorpusReader.raw -nltk.corpus.reader.cmudict.CorpusReader.readme -nltk.corpus.reader.cmudict.CorpusReader.root -nltk.corpus.reader.cmudict.StreamBackedCorpusView.__add__ -nltk.corpus.reader.cmudict.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.cmudict.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.cmudict.StreamBackedCorpusView.__len__ -nltk.corpus.reader.cmudict.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.cmudict.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.cmudict.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.cmudict.StreamBackedCorpusView.fileid -nltk.corpus.reader.cmudict.StreamBackedCorpusView.read_block -nltk.corpus.reader.cmudict.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.cmudict.SyntaxCorpusReader.sents -nltk.corpus.reader.comparative_sents.CategorizedCorpusReader.paras -nltk.corpus.reader.comparative_sents.CategorizedCorpusReader.raw -nltk.corpus.reader.comparative_sents.CategorizedCorpusReader.sents -nltk.corpus.reader.comparative_sents.ComparativeSentencesCorpusReader.CorpusView -nltk.corpus.reader.comparative_sents.CorpusReader.citation -nltk.corpus.reader.comparative_sents.CorpusReader.license -nltk.corpus.reader.comparative_sents.CorpusReader.raw -nltk.corpus.reader.comparative_sents.CorpusReader.readme -nltk.corpus.reader.comparative_sents.CorpusReader.root -nltk.corpus.reader.comparative_sents.LineTokenizer -nltk.corpus.reader.comparative_sents.LineTokenizer.span_tokenize -nltk.corpus.reader.comparative_sents.LineTokenizer.tokenize -nltk.corpus.reader.comparative_sents.StanfordSegmenter -nltk.corpus.reader.comparative_sents.StreamBackedCorpusView.__add__ -nltk.corpus.reader.comparative_sents.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.comparative_sents.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.comparative_sents.StreamBackedCorpusView.__len__ -nltk.corpus.reader.comparative_sents.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.comparative_sents.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.comparative_sents.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.comparative_sents.StreamBackedCorpusView.fileid -nltk.corpus.reader.comparative_sents.StreamBackedCorpusView.read_block -nltk.corpus.reader.comparative_sents.SyllableTokenizer -nltk.corpus.reader.comparative_sents.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.comparative_sents.SyntaxCorpusReader.sents -nltk.corpus.reader.comparative_sents.TreebankWordTokenizer -nltk.corpus.reader.comparative_sents.TreebankWordTokenizer.CONTRACTIONS2 -nltk.corpus.reader.comparative_sents.TreebankWordTokenizer.CONTRACTIONS3 -nltk.corpus.reader.comparative_sents.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.corpus.reader.comparative_sents.TreebankWordTokenizer.DOUBLE_DASHES -nltk.corpus.reader.comparative_sents.TreebankWordTokenizer.ENDING_QUOTES -nltk.corpus.reader.comparative_sents.TreebankWordTokenizer.PARENS_BRACKETS -nltk.corpus.reader.comparative_sents.TreebankWordTokenizer.PUNCTUATION -nltk.corpus.reader.comparative_sents.TreebankWordTokenizer.STARTING_QUOTES -nltk.corpus.reader.conll.CategorizedCorpusReader.paras -nltk.corpus.reader.conll.CategorizedCorpusReader.raw -nltk.corpus.reader.conll.CategorizedCorpusReader.sents -nltk.corpus.reader.conll.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.conll.ConcatenatedCorpusView.close -nltk.corpus.reader.conll.CorpusReader.citation -nltk.corpus.reader.conll.CorpusReader.license -nltk.corpus.reader.conll.CorpusReader.raw -nltk.corpus.reader.conll.CorpusReader.readme -nltk.corpus.reader.conll.CorpusReader.root -nltk.corpus.reader.conll.StreamBackedCorpusView.__add__ -nltk.corpus.reader.conll.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.conll.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.conll.StreamBackedCorpusView.__len__ -nltk.corpus.reader.conll.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.conll.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.conll.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.conll.StreamBackedCorpusView.fileid -nltk.corpus.reader.conll.StreamBackedCorpusView.read_block -nltk.corpus.reader.conll.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.conll.SyntaxCorpusReader.sents -nltk.corpus.reader.crubadan.CorpusReader.citation -nltk.corpus.reader.crubadan.CorpusReader.license -nltk.corpus.reader.crubadan.CorpusReader.raw -nltk.corpus.reader.crubadan.CorpusReader.readme -nltk.corpus.reader.crubadan.CorpusReader.root -nltk.corpus.reader.dependency.DependencyCorpusReader.__init__ -nltk.corpus.reader.dependency.DependencyCorpusReader.tagged_sents -nltk.corpus.reader.dependency.DependencyCorpusReader.tagged_words -nltk.corpus.reader.dependency.DependencyCorpusReader.words -nltk.corpus.reader.framenet_typed -nltk.corpus.reader.ieer.CategorizedCorpusReader.paras -nltk.corpus.reader.ieer.CategorizedCorpusReader.raw -nltk.corpus.reader.ieer.CategorizedCorpusReader.sents -nltk.corpus.reader.ieer.CorpusReader.citation -nltk.corpus.reader.ieer.CorpusReader.license -nltk.corpus.reader.ieer.CorpusReader.raw -nltk.corpus.reader.ieer.CorpusReader.readme -nltk.corpus.reader.ieer.CorpusReader.root -nltk.corpus.reader.ieer.StreamBackedCorpusView.__add__ -nltk.corpus.reader.ieer.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.ieer.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.ieer.StreamBackedCorpusView.__len__ -nltk.corpus.reader.ieer.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.ieer.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.ieer.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.ieer.StreamBackedCorpusView.fileid -nltk.corpus.reader.ieer.StreamBackedCorpusView.read_block -nltk.corpus.reader.ieer.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.ieer.SyntaxCorpusReader.sents -nltk.corpus.reader.indian.CategorizedCorpusReader.paras -nltk.corpus.reader.indian.CategorizedCorpusReader.raw -nltk.corpus.reader.indian.CategorizedCorpusReader.sents -nltk.corpus.reader.indian.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.indian.ConcatenatedCorpusView.close -nltk.corpus.reader.indian.CorpusReader.citation -nltk.corpus.reader.indian.CorpusReader.license -nltk.corpus.reader.indian.CorpusReader.raw -nltk.corpus.reader.indian.CorpusReader.readme -nltk.corpus.reader.indian.CorpusReader.root -nltk.corpus.reader.indian.StreamBackedCorpusView.__add__ -nltk.corpus.reader.indian.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.indian.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.indian.StreamBackedCorpusView.__len__ -nltk.corpus.reader.indian.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.indian.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.indian.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.indian.StreamBackedCorpusView.fileid -nltk.corpus.reader.indian.StreamBackedCorpusView.read_block -nltk.corpus.reader.indian.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.indian.SyntaxCorpusReader.sents -nltk.corpus.reader.ipipan.IPIPANCorpusReader -nltk.corpus.reader.ipipan.IPIPANCorpusView -nltk.corpus.reader.knbc.CorpusReader.citation -nltk.corpus.reader.knbc.CorpusReader.license -nltk.corpus.reader.knbc.CorpusReader.raw -nltk.corpus.reader.knbc.CorpusReader.readme -nltk.corpus.reader.knbc.CorpusReader.root -nltk.corpus.reader.knbc.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.knbc.SyntaxCorpusReader.sents -nltk.corpus.reader.lin.CorpusReader.citation -nltk.corpus.reader.lin.CorpusReader.license -nltk.corpus.reader.lin.CorpusReader.raw -nltk.corpus.reader.lin.CorpusReader.readme -nltk.corpus.reader.lin.CorpusReader.root -nltk.corpus.reader.markdown.CategorizedCorpusReader.paras -nltk.corpus.reader.markdown.CategorizedCorpusReader.raw -nltk.corpus.reader.markdown.CategorizedCorpusReader.sents -nltk.corpus.reader.markdown.PlaintextCorpusReader.CorpusView -nltk.corpus.reader.mte_typed -nltk.corpus.reader.nkjp.NKJPCorpusReader -nltk.corpus.reader.nkjp.NKJPCorpus_Header_View -nltk.corpus.reader.nkjp.NKJPCorpus_Morph_View -nltk.corpus.reader.nkjp.NKJPCorpus_Segmentation_View -nltk.corpus.reader.nkjp.NKJPCorpus_Text_View -nltk.corpus.reader.nkjp.XML_Tool -nltk.corpus.reader.nombank.NombankChainTreePointer -nltk.corpus.reader.nombank.NombankCorpusReader.instances -nltk.corpus.reader.nombank.NombankCorpusReader.lines -nltk.corpus.reader.nombank.NombankCorpusReader.roleset -nltk.corpus.reader.nombank.NombankCorpusReader.rolesets -nltk.corpus.reader.nombank.NombankInstance -nltk.corpus.reader.nombank.NombankPointer -nltk.corpus.reader.nombank.NombankSplitTreePointer -nltk.corpus.reader.nombank.NombankTreePointer -nltk.corpus.reader.norm -nltk.corpus.reader.nps_chat.CategorizedCorpusReader.paras -nltk.corpus.reader.nps_chat.CategorizedCorpusReader.raw -nltk.corpus.reader.nps_chat.CategorizedCorpusReader.sents -nltk.corpus.reader.nps_chat.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.nps_chat.ConcatenatedCorpusView.close -nltk.corpus.reader.nps_chat.CorpusReader.citation -nltk.corpus.reader.nps_chat.CorpusReader.license -nltk.corpus.reader.nps_chat.CorpusReader.raw -nltk.corpus.reader.nps_chat.CorpusReader.readme -nltk.corpus.reader.nps_chat.CorpusReader.root -nltk.corpus.reader.nps_chat.StreamBackedCorpusView.__add__ -nltk.corpus.reader.nps_chat.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.nps_chat.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.nps_chat.StreamBackedCorpusView.__len__ -nltk.corpus.reader.nps_chat.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.nps_chat.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.nps_chat.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.nps_chat.StreamBackedCorpusView.fileid -nltk.corpus.reader.nps_chat.StreamBackedCorpusView.read_block -nltk.corpus.reader.nps_chat.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.nps_chat.SyntaxCorpusReader.sents -nltk.corpus.reader.opinion_lexicon.CategorizedCorpusReader.paras -nltk.corpus.reader.opinion_lexicon.CategorizedCorpusReader.raw -nltk.corpus.reader.opinion_lexicon.CategorizedCorpusReader.sents -nltk.corpus.reader.opinion_lexicon.CorpusReader.citation -nltk.corpus.reader.opinion_lexicon.CorpusReader.license -nltk.corpus.reader.opinion_lexicon.CorpusReader.raw -nltk.corpus.reader.opinion_lexicon.CorpusReader.readme -nltk.corpus.reader.opinion_lexicon.CorpusReader.root -nltk.corpus.reader.opinion_lexicon.OpinionLexiconCorpusReader.CorpusView -nltk.corpus.reader.opinion_lexicon.StreamBackedCorpusView.__add__ -nltk.corpus.reader.opinion_lexicon.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.opinion_lexicon.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.opinion_lexicon.StreamBackedCorpusView.__len__ -nltk.corpus.reader.opinion_lexicon.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.opinion_lexicon.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.opinion_lexicon.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.opinion_lexicon.StreamBackedCorpusView.fileid -nltk.corpus.reader.opinion_lexicon.StreamBackedCorpusView.read_block -nltk.corpus.reader.opinion_lexicon.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.opinion_lexicon.SyntaxCorpusReader.sents -nltk.corpus.reader.panlex_lite.CorpusReader.citation -nltk.corpus.reader.panlex_lite.CorpusReader.license -nltk.corpus.reader.panlex_lite.CorpusReader.raw -nltk.corpus.reader.panlex_lite.CorpusReader.readme -nltk.corpus.reader.panlex_lite.CorpusReader.root -nltk.corpus.reader.panlex_swadesh.CategorizedCorpusReader.paras -nltk.corpus.reader.panlex_swadesh.CategorizedCorpusReader.raw -nltk.corpus.reader.panlex_swadesh.CategorizedCorpusReader.sents -nltk.corpus.reader.panlex_swadesh.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.panlex_swadesh.ConcatenatedCorpusView.close -nltk.corpus.reader.panlex_swadesh.CorpusReader.citation -nltk.corpus.reader.panlex_swadesh.CorpusReader.license -nltk.corpus.reader.panlex_swadesh.CorpusReader.raw -nltk.corpus.reader.panlex_swadesh.CorpusReader.readme -nltk.corpus.reader.panlex_swadesh.CorpusReader.root -nltk.corpus.reader.panlex_swadesh.StreamBackedCorpusView.__add__ -nltk.corpus.reader.panlex_swadesh.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.panlex_swadesh.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.panlex_swadesh.StreamBackedCorpusView.__len__ -nltk.corpus.reader.panlex_swadesh.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.panlex_swadesh.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.panlex_swadesh.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.panlex_swadesh.StreamBackedCorpusView.fileid -nltk.corpus.reader.panlex_swadesh.StreamBackedCorpusView.read_block -nltk.corpus.reader.panlex_swadesh.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.panlex_swadesh.SyntaxCorpusReader.sents -nltk.corpus.reader.pl196x.CategorizedCorpusReader.paras -nltk.corpus.reader.pl196x.CategorizedCorpusReader.raw -nltk.corpus.reader.pl196x.CategorizedCorpusReader.sents -nltk.corpus.reader.pl196x.CorpusReader.citation -nltk.corpus.reader.pl196x.CorpusReader.license -nltk.corpus.reader.pl196x.CorpusReader.raw -nltk.corpus.reader.pl196x.CorpusReader.readme -nltk.corpus.reader.pl196x.CorpusReader.root -nltk.corpus.reader.pl196x.StreamBackedCorpusView.__add__ -nltk.corpus.reader.pl196x.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.pl196x.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.pl196x.StreamBackedCorpusView.__len__ -nltk.corpus.reader.pl196x.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.pl196x.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.pl196x.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.pl196x.StreamBackedCorpusView.fileid -nltk.corpus.reader.pl196x.StreamBackedCorpusView.read_block -nltk.corpus.reader.pl196x.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.pl196x.SyntaxCorpusReader.sents -nltk.corpus.reader.plaintext.CategorizedCorpusReader.paras -nltk.corpus.reader.plaintext.CategorizedCorpusReader.raw -nltk.corpus.reader.plaintext.CategorizedCorpusReader.sents -nltk.corpus.reader.plaintext.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.plaintext.ConcatenatedCorpusView.close -nltk.corpus.reader.plaintext.CorpusReader.citation -nltk.corpus.reader.plaintext.CorpusReader.license -nltk.corpus.reader.plaintext.CorpusReader.raw -nltk.corpus.reader.plaintext.CorpusReader.readme -nltk.corpus.reader.plaintext.CorpusReader.root -nltk.corpus.reader.plaintext.LineTokenizer -nltk.corpus.reader.plaintext.LineTokenizer.span_tokenize -nltk.corpus.reader.plaintext.LineTokenizer.tokenize -nltk.corpus.reader.plaintext.PlaintextCorpusReader.CorpusView -nltk.corpus.reader.plaintext.StanfordSegmenter -nltk.corpus.reader.plaintext.StreamBackedCorpusView.__add__ -nltk.corpus.reader.plaintext.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.plaintext.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.plaintext.StreamBackedCorpusView.__len__ -nltk.corpus.reader.plaintext.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.plaintext.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.plaintext.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.plaintext.StreamBackedCorpusView.fileid -nltk.corpus.reader.plaintext.StreamBackedCorpusView.read_block -nltk.corpus.reader.plaintext.SyllableTokenizer -nltk.corpus.reader.plaintext.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.plaintext.SyntaxCorpusReader.sents -nltk.corpus.reader.plaintext.TreebankWordTokenizer -nltk.corpus.reader.plaintext.TreebankWordTokenizer.CONTRACTIONS2 -nltk.corpus.reader.plaintext.TreebankWordTokenizer.CONTRACTIONS3 -nltk.corpus.reader.plaintext.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.corpus.reader.plaintext.TreebankWordTokenizer.DOUBLE_DASHES -nltk.corpus.reader.plaintext.TreebankWordTokenizer.ENDING_QUOTES -nltk.corpus.reader.plaintext.TreebankWordTokenizer.PARENS_BRACKETS -nltk.corpus.reader.plaintext.TreebankWordTokenizer.PUNCTUATION -nltk.corpus.reader.plaintext.TreebankWordTokenizer.STARTING_QUOTES -nltk.corpus.reader.ppattach.CategorizedCorpusReader.paras -nltk.corpus.reader.ppattach.CategorizedCorpusReader.raw -nltk.corpus.reader.ppattach.CategorizedCorpusReader.sents -nltk.corpus.reader.ppattach.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.ppattach.ConcatenatedCorpusView.close -nltk.corpus.reader.ppattach.CorpusReader.citation -nltk.corpus.reader.ppattach.CorpusReader.license -nltk.corpus.reader.ppattach.CorpusReader.raw -nltk.corpus.reader.ppattach.CorpusReader.readme -nltk.corpus.reader.ppattach.CorpusReader.root -nltk.corpus.reader.ppattach.StreamBackedCorpusView.__add__ -nltk.corpus.reader.ppattach.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.ppattach.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.ppattach.StreamBackedCorpusView.__len__ -nltk.corpus.reader.ppattach.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.ppattach.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.ppattach.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.ppattach.StreamBackedCorpusView.fileid -nltk.corpus.reader.ppattach.StreamBackedCorpusView.read_block -nltk.corpus.reader.ppattach.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.ppattach.SyntaxCorpusReader.sents -nltk.corpus.reader.propbank.CategorizedCorpusReader.paras -nltk.corpus.reader.propbank.CategorizedCorpusReader.raw -nltk.corpus.reader.propbank.CategorizedCorpusReader.sents -nltk.corpus.reader.propbank.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.propbank.ConcatenatedCorpusView.close -nltk.corpus.reader.propbank.CorpusReader.citation -nltk.corpus.reader.propbank.CorpusReader.license -nltk.corpus.reader.propbank.CorpusReader.raw -nltk.corpus.reader.propbank.CorpusReader.readme -nltk.corpus.reader.propbank.CorpusReader.root -nltk.corpus.reader.propbank.PropbankInstance.tree -nltk.corpus.reader.propbank.PropbankTreePointer.__ge__ -nltk.corpus.reader.propbank.PropbankTreePointer.__gt__ -nltk.corpus.reader.propbank.PropbankTreePointer.__le__ -nltk.corpus.reader.propbank.StreamBackedCorpusView.__add__ -nltk.corpus.reader.propbank.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.propbank.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.propbank.StreamBackedCorpusView.__len__ -nltk.corpus.reader.propbank.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.propbank.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.propbank.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.propbank.StreamBackedCorpusView.fileid -nltk.corpus.reader.propbank.StreamBackedCorpusView.read_block -nltk.corpus.reader.propbank.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.propbank.SyntaxCorpusReader.sents -nltk.corpus.reader.pros_cons.CategorizedCorpusReader.paras -nltk.corpus.reader.pros_cons.CategorizedCorpusReader.raw -nltk.corpus.reader.pros_cons.CategorizedCorpusReader.sents -nltk.corpus.reader.pros_cons.CorpusReader.citation -nltk.corpus.reader.pros_cons.CorpusReader.license -nltk.corpus.reader.pros_cons.CorpusReader.raw -nltk.corpus.reader.pros_cons.CorpusReader.readme -nltk.corpus.reader.pros_cons.CorpusReader.root -nltk.corpus.reader.pros_cons.LineTokenizer -nltk.corpus.reader.pros_cons.LineTokenizer.span_tokenize -nltk.corpus.reader.pros_cons.LineTokenizer.tokenize -nltk.corpus.reader.pros_cons.ProsConsCorpusReader.CorpusView -nltk.corpus.reader.pros_cons.StanfordSegmenter -nltk.corpus.reader.pros_cons.StreamBackedCorpusView.__add__ -nltk.corpus.reader.pros_cons.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.pros_cons.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.pros_cons.StreamBackedCorpusView.__len__ -nltk.corpus.reader.pros_cons.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.pros_cons.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.pros_cons.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.pros_cons.StreamBackedCorpusView.fileid -nltk.corpus.reader.pros_cons.StreamBackedCorpusView.read_block -nltk.corpus.reader.pros_cons.SyllableTokenizer -nltk.corpus.reader.pros_cons.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.pros_cons.SyntaxCorpusReader.sents -nltk.corpus.reader.pros_cons.TreebankWordTokenizer -nltk.corpus.reader.pros_cons.TreebankWordTokenizer.CONTRACTIONS2 -nltk.corpus.reader.pros_cons.TreebankWordTokenizer.CONTRACTIONS3 -nltk.corpus.reader.pros_cons.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.corpus.reader.pros_cons.TreebankWordTokenizer.DOUBLE_DASHES -nltk.corpus.reader.pros_cons.TreebankWordTokenizer.ENDING_QUOTES -nltk.corpus.reader.pros_cons.TreebankWordTokenizer.PARENS_BRACKETS -nltk.corpus.reader.pros_cons.TreebankWordTokenizer.PUNCTUATION -nltk.corpus.reader.pros_cons.TreebankWordTokenizer.STARTING_QUOTES -nltk.corpus.reader.reviews.CategorizedCorpusReader.paras -nltk.corpus.reader.reviews.CategorizedCorpusReader.raw -nltk.corpus.reader.reviews.CategorizedCorpusReader.sents -nltk.corpus.reader.reviews.CorpusReader.citation -nltk.corpus.reader.reviews.CorpusReader.license -nltk.corpus.reader.reviews.CorpusReader.raw -nltk.corpus.reader.reviews.CorpusReader.readme -nltk.corpus.reader.reviews.CorpusReader.root -nltk.corpus.reader.reviews.LineTokenizer -nltk.corpus.reader.reviews.LineTokenizer.span_tokenize -nltk.corpus.reader.reviews.LineTokenizer.tokenize -nltk.corpus.reader.reviews.ReviewsCorpusReader.CorpusView -nltk.corpus.reader.reviews.StanfordSegmenter -nltk.corpus.reader.reviews.StreamBackedCorpusView.__add__ -nltk.corpus.reader.reviews.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.reviews.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.reviews.StreamBackedCorpusView.__len__ -nltk.corpus.reader.reviews.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.reviews.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.reviews.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.reviews.StreamBackedCorpusView.fileid -nltk.corpus.reader.reviews.StreamBackedCorpusView.read_block -nltk.corpus.reader.reviews.SyllableTokenizer -nltk.corpus.reader.reviews.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.reviews.SyntaxCorpusReader.sents -nltk.corpus.reader.reviews.TreebankWordTokenizer -nltk.corpus.reader.reviews.TreebankWordTokenizer.CONTRACTIONS2 -nltk.corpus.reader.reviews.TreebankWordTokenizer.CONTRACTIONS3 -nltk.corpus.reader.reviews.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.corpus.reader.reviews.TreebankWordTokenizer.DOUBLE_DASHES -nltk.corpus.reader.reviews.TreebankWordTokenizer.ENDING_QUOTES -nltk.corpus.reader.reviews.TreebankWordTokenizer.PARENS_BRACKETS -nltk.corpus.reader.reviews.TreebankWordTokenizer.PUNCTUATION -nltk.corpus.reader.reviews.TreebankWordTokenizer.STARTING_QUOTES -nltk.corpus.reader.rte.CategorizedCorpusReader.paras -nltk.corpus.reader.rte.CategorizedCorpusReader.raw -nltk.corpus.reader.rte.CategorizedCorpusReader.sents -nltk.corpus.reader.rte.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.rte.ConcatenatedCorpusView.close -nltk.corpus.reader.rte.CorpusReader.citation -nltk.corpus.reader.rte.CorpusReader.license -nltk.corpus.reader.rte.CorpusReader.raw -nltk.corpus.reader.rte.CorpusReader.readme -nltk.corpus.reader.rte.CorpusReader.root -nltk.corpus.reader.rte.StreamBackedCorpusView.__add__ -nltk.corpus.reader.rte.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.rte.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.rte.StreamBackedCorpusView.__len__ -nltk.corpus.reader.rte.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.rte.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.rte.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.rte.StreamBackedCorpusView.fileid -nltk.corpus.reader.rte.StreamBackedCorpusView.read_block -nltk.corpus.reader.rte.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.rte.SyntaxCorpusReader.sents -nltk.corpus.reader.rte.norm -nltk.corpus.reader.semcor.CategorizedCorpusReader.paras -nltk.corpus.reader.semcor.CategorizedCorpusReader.raw -nltk.corpus.reader.semcor.CategorizedCorpusReader.sents -nltk.corpus.reader.semcor.CorpusReader.citation -nltk.corpus.reader.semcor.CorpusReader.license -nltk.corpus.reader.semcor.CorpusReader.raw -nltk.corpus.reader.semcor.CorpusReader.readme -nltk.corpus.reader.semcor.CorpusReader.root -nltk.corpus.reader.semcor.StreamBackedCorpusView.__add__ -nltk.corpus.reader.semcor.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.semcor.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.semcor.StreamBackedCorpusView.__len__ -nltk.corpus.reader.semcor.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.semcor.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.semcor.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.semcor.StreamBackedCorpusView.fileid -nltk.corpus.reader.semcor.StreamBackedCorpusView.read_block -nltk.corpus.reader.semcor.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.semcor.SyntaxCorpusReader.sents -nltk.corpus.reader.senseval.CategorizedCorpusReader.paras -nltk.corpus.reader.senseval.CategorizedCorpusReader.raw -nltk.corpus.reader.senseval.CategorizedCorpusReader.sents -nltk.corpus.reader.senseval.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.senseval.ConcatenatedCorpusView.close -nltk.corpus.reader.senseval.CorpusReader.citation -nltk.corpus.reader.senseval.CorpusReader.license -nltk.corpus.reader.senseval.CorpusReader.raw -nltk.corpus.reader.senseval.CorpusReader.readme -nltk.corpus.reader.senseval.CorpusReader.root -nltk.corpus.reader.senseval.LineTokenizer -nltk.corpus.reader.senseval.LineTokenizer.span_tokenize -nltk.corpus.reader.senseval.LineTokenizer.tokenize -nltk.corpus.reader.senseval.StanfordSegmenter -nltk.corpus.reader.senseval.StreamBackedCorpusView.__add__ -nltk.corpus.reader.senseval.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.senseval.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.senseval.StreamBackedCorpusView.__len__ -nltk.corpus.reader.senseval.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.senseval.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.senseval.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.senseval.StreamBackedCorpusView.fileid -nltk.corpus.reader.senseval.StreamBackedCorpusView.read_block -nltk.corpus.reader.senseval.SyllableTokenizer -nltk.corpus.reader.senseval.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.senseval.SyntaxCorpusReader.sents -nltk.corpus.reader.senseval.TreebankWordTokenizer -nltk.corpus.reader.senseval.TreebankWordTokenizer.CONTRACTIONS2 -nltk.corpus.reader.senseval.TreebankWordTokenizer.CONTRACTIONS3 -nltk.corpus.reader.senseval.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.corpus.reader.senseval.TreebankWordTokenizer.DOUBLE_DASHES -nltk.corpus.reader.senseval.TreebankWordTokenizer.ENDING_QUOTES -nltk.corpus.reader.senseval.TreebankWordTokenizer.PARENS_BRACKETS -nltk.corpus.reader.senseval.TreebankWordTokenizer.PUNCTUATION -nltk.corpus.reader.senseval.TreebankWordTokenizer.STARTING_QUOTES -nltk.corpus.reader.sentiwordnet.CorpusReader.citation -nltk.corpus.reader.sentiwordnet.CorpusReader.license -nltk.corpus.reader.sentiwordnet.CorpusReader.raw -nltk.corpus.reader.sentiwordnet.CorpusReader.readme -nltk.corpus.reader.sentiwordnet.CorpusReader.root -nltk.corpus.reader.sinica_treebank.CategorizedCorpusReader.paras -nltk.corpus.reader.sinica_treebank.CategorizedCorpusReader.raw -nltk.corpus.reader.sinica_treebank.CategorizedCorpusReader.sents -nltk.corpus.reader.sinica_treebank.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.sinica_treebank.ConcatenatedCorpusView.close -nltk.corpus.reader.sinica_treebank.CorpusReader.citation -nltk.corpus.reader.sinica_treebank.CorpusReader.license -nltk.corpus.reader.sinica_treebank.CorpusReader.raw -nltk.corpus.reader.sinica_treebank.CorpusReader.readme -nltk.corpus.reader.sinica_treebank.CorpusReader.root -nltk.corpus.reader.sinica_treebank.StreamBackedCorpusView.__add__ -nltk.corpus.reader.sinica_treebank.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.sinica_treebank.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.sinica_treebank.StreamBackedCorpusView.__len__ -nltk.corpus.reader.sinica_treebank.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.sinica_treebank.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.sinica_treebank.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.sinica_treebank.StreamBackedCorpusView.fileid -nltk.corpus.reader.sinica_treebank.StreamBackedCorpusView.read_block -nltk.corpus.reader.sinica_treebank.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.sinica_treebank.SyntaxCorpusReader.sents -nltk.corpus.reader.string_category.CategorizedCorpusReader.paras -nltk.corpus.reader.string_category.CategorizedCorpusReader.raw -nltk.corpus.reader.string_category.CategorizedCorpusReader.sents -nltk.corpus.reader.string_category.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.string_category.ConcatenatedCorpusView.close -nltk.corpus.reader.string_category.CorpusReader.citation -nltk.corpus.reader.string_category.CorpusReader.license -nltk.corpus.reader.string_category.CorpusReader.raw -nltk.corpus.reader.string_category.CorpusReader.readme -nltk.corpus.reader.string_category.CorpusReader.root -nltk.corpus.reader.string_category.StreamBackedCorpusView.__add__ -nltk.corpus.reader.string_category.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.string_category.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.string_category.StreamBackedCorpusView.__len__ -nltk.corpus.reader.string_category.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.string_category.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.string_category.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.string_category.StreamBackedCorpusView.fileid -nltk.corpus.reader.string_category.StreamBackedCorpusView.read_block -nltk.corpus.reader.string_category.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.string_category.SyntaxCorpusReader.sents -nltk.corpus.reader.switchboard.CategorizedCorpusReader.paras -nltk.corpus.reader.switchboard.CategorizedCorpusReader.raw -nltk.corpus.reader.switchboard.CategorizedCorpusReader.sents -nltk.corpus.reader.switchboard.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.switchboard.ConcatenatedCorpusView.close -nltk.corpus.reader.switchboard.CorpusReader.citation -nltk.corpus.reader.switchboard.CorpusReader.license -nltk.corpus.reader.switchboard.CorpusReader.raw -nltk.corpus.reader.switchboard.CorpusReader.readme -nltk.corpus.reader.switchboard.CorpusReader.root -nltk.corpus.reader.switchboard.StreamBackedCorpusView.__add__ -nltk.corpus.reader.switchboard.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.switchboard.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.switchboard.StreamBackedCorpusView.__len__ -nltk.corpus.reader.switchboard.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.switchboard.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.switchboard.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.switchboard.StreamBackedCorpusView.fileid -nltk.corpus.reader.switchboard.StreamBackedCorpusView.read_block -nltk.corpus.reader.switchboard.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.switchboard.SyntaxCorpusReader.sents -nltk.corpus.reader.tagged.CategorizedCorpusReader.paras -nltk.corpus.reader.tagged.CategorizedCorpusReader.raw -nltk.corpus.reader.tagged.CategorizedCorpusReader.sents -nltk.corpus.reader.tagged.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.tagged.ConcatenatedCorpusView.close -nltk.corpus.reader.tagged.CorpusReader.citation -nltk.corpus.reader.tagged.CorpusReader.license -nltk.corpus.reader.tagged.CorpusReader.raw -nltk.corpus.reader.tagged.CorpusReader.readme -nltk.corpus.reader.tagged.CorpusReader.root -nltk.corpus.reader.tagged.LineTokenizer -nltk.corpus.reader.tagged.LineTokenizer.span_tokenize -nltk.corpus.reader.tagged.LineTokenizer.tokenize -nltk.corpus.reader.tagged.StanfordSegmenter -nltk.corpus.reader.tagged.StreamBackedCorpusView.__add__ -nltk.corpus.reader.tagged.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.tagged.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.tagged.StreamBackedCorpusView.__len__ -nltk.corpus.reader.tagged.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.tagged.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.tagged.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.tagged.StreamBackedCorpusView.fileid -nltk.corpus.reader.tagged.StreamBackedCorpusView.read_block -nltk.corpus.reader.tagged.SyllableTokenizer -nltk.corpus.reader.tagged.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.tagged.SyntaxCorpusReader.sents -nltk.corpus.reader.tagged.TreebankWordTokenizer -nltk.corpus.reader.tagged.TreebankWordTokenizer.CONTRACTIONS2 -nltk.corpus.reader.tagged.TreebankWordTokenizer.CONTRACTIONS3 -nltk.corpus.reader.tagged.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.corpus.reader.tagged.TreebankWordTokenizer.DOUBLE_DASHES -nltk.corpus.reader.tagged.TreebankWordTokenizer.ENDING_QUOTES -nltk.corpus.reader.tagged.TreebankWordTokenizer.PARENS_BRACKETS -nltk.corpus.reader.tagged.TreebankWordTokenizer.PUNCTUATION -nltk.corpus.reader.tagged.TreebankWordTokenizer.STARTING_QUOTES +nltk.corpus.reader.api.SyntaxCorpusReader +nltk.corpus.reader.bcp47 +nltk.corpus.reader.bnc +nltk.corpus.reader.bracket_parse +nltk.corpus.reader.categorized_sents +nltk.corpus.reader.chasen +nltk.corpus.reader.childes +nltk.corpus.reader.chunked +nltk.corpus.reader.cmudict +nltk.corpus.reader.comparative_sents +nltk.corpus.reader.conll +nltk.corpus.reader.crubadan +nltk.corpus.reader.dependency +nltk.corpus.reader.find_corpus_fileids +nltk.corpus.reader.framenet +nltk.corpus.reader.ieer +nltk.corpus.reader.indian +nltk.corpus.reader.ipipan +nltk.corpus.reader.knbc +nltk.corpus.reader.lin +nltk.corpus.reader.markdown +nltk.corpus.reader.mte +nltk.corpus.reader.nkjp +nltk.corpus.reader.nombank +nltk.corpus.reader.nps_chat +nltk.corpus.reader.opinion_lexicon +nltk.corpus.reader.panlex_lite +nltk.corpus.reader.panlex_swadesh +nltk.corpus.reader.pl196x +nltk.corpus.reader.plaintext +nltk.corpus.reader.ppattach +nltk.corpus.reader.propbank +nltk.corpus.reader.pros_cons +nltk.corpus.reader.reviews +nltk.corpus.reader.rte +nltk.corpus.reader.semcor +nltk.corpus.reader.senseval +nltk.corpus.reader.sentiwordnet +nltk.corpus.reader.sinica_treebank +nltk.corpus.reader.string_category +nltk.corpus.reader.switchboard +nltk.corpus.reader.tagged nltk.corpus.reader.tagged_treebank_para_block_reader -nltk.corpus.reader.timit.CategorizedCorpusReader.paras -nltk.corpus.reader.timit.CategorizedCorpusReader.raw -nltk.corpus.reader.timit.CategorizedCorpusReader.sents -nltk.corpus.reader.timit.CorpusReader.citation -nltk.corpus.reader.timit.CorpusReader.license -nltk.corpus.reader.timit.CorpusReader.raw -nltk.corpus.reader.timit.CorpusReader.readme -nltk.corpus.reader.timit.CorpusReader.root -nltk.corpus.reader.timit.StreamBackedCorpusView.__add__ -nltk.corpus.reader.timit.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.timit.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.timit.StreamBackedCorpusView.__len__ -nltk.corpus.reader.timit.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.timit.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.timit.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.timit.StreamBackedCorpusView.fileid -nltk.corpus.reader.timit.StreamBackedCorpusView.read_block -nltk.corpus.reader.timit.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.timit.SyntaxCorpusReader.sents -nltk.corpus.reader.toolbox.CategorizedCorpusReader.paras -nltk.corpus.reader.toolbox.CategorizedCorpusReader.raw -nltk.corpus.reader.toolbox.CategorizedCorpusReader.sents -nltk.corpus.reader.toolbox.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.toolbox.ConcatenatedCorpusView.close -nltk.corpus.reader.toolbox.CorpusReader.citation -nltk.corpus.reader.toolbox.CorpusReader.license -nltk.corpus.reader.toolbox.CorpusReader.raw -nltk.corpus.reader.toolbox.CorpusReader.readme -nltk.corpus.reader.toolbox.CorpusReader.root -nltk.corpus.reader.toolbox.StreamBackedCorpusView.__add__ -nltk.corpus.reader.toolbox.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.toolbox.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.toolbox.StreamBackedCorpusView.__len__ -nltk.corpus.reader.toolbox.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.toolbox.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.toolbox.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.toolbox.StreamBackedCorpusView.fileid -nltk.corpus.reader.toolbox.StreamBackedCorpusView.read_block -nltk.corpus.reader.toolbox.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.toolbox.SyntaxCorpusReader.sents -nltk.corpus.reader.twitter.CorpusReader.citation -nltk.corpus.reader.twitter.CorpusReader.license -nltk.corpus.reader.twitter.CorpusReader.raw -nltk.corpus.reader.twitter.CorpusReader.readme -nltk.corpus.reader.twitter.CorpusReader.root -nltk.corpus.reader.twitter.StreamBackedCorpusView.__add__ -nltk.corpus.reader.twitter.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.twitter.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.twitter.StreamBackedCorpusView.__len__ -nltk.corpus.reader.twitter.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.twitter.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.twitter.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.twitter.StreamBackedCorpusView.fileid -nltk.corpus.reader.twitter.StreamBackedCorpusView.read_block -nltk.corpus.reader.twitter.TwitterCorpusReader.CorpusView -nltk.corpus.reader.udhr.PlaintextCorpusReader.CorpusView -nltk.corpus.reader.util.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.util.ConcatenatedCorpusView.close +nltk.corpus.reader.timit +nltk.corpus.reader.toolbox +nltk.corpus.reader.twitter +nltk.corpus.reader.udhr +nltk.corpus.reader.util.ConcatenatedCorpusView nltk.corpus.reader.util.StreamBackedCorpusView.__add__ nltk.corpus.reader.util.StreamBackedCorpusView.__enter__ nltk.corpus.reader.util.StreamBackedCorpusView.__exit__ @@ -1379,41 +363,21 @@ nltk.corpus.reader.util.StreamBackedCorpusView.__radd__ nltk.corpus.reader.util.StreamBackedCorpusView.__rmul__ nltk.corpus.reader.util.StreamBackedCorpusView.fileid nltk.corpus.reader.util.StreamBackedCorpusView.read_block +nltk.corpus.reader.util.concat +nltk.corpus.reader.util.find_corpus_fileids nltk.corpus.reader.util.read_alignedsent_block +nltk.corpus.reader.util.read_blankline_block nltk.corpus.reader.util.read_line_block +nltk.corpus.reader.util.read_regexp_block nltk.corpus.reader.util.read_sexpr_block nltk.corpus.reader.util.read_whitespace_block nltk.corpus.reader.util.read_wordpunct_block nltk.corpus.reader.util.tagged_treebank_para_block_reader -nltk.corpus.reader.wordlist.CategorizedCorpusReader.paras -nltk.corpus.reader.wordlist.CategorizedCorpusReader.raw -nltk.corpus.reader.wordlist.CategorizedCorpusReader.sents -nltk.corpus.reader.wordlist.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.wordlist.ConcatenatedCorpusView.close -nltk.corpus.reader.wordlist.CorpusReader.citation -nltk.corpus.reader.wordlist.CorpusReader.license -nltk.corpus.reader.wordlist.CorpusReader.raw -nltk.corpus.reader.wordlist.CorpusReader.readme -nltk.corpus.reader.wordlist.CorpusReader.root -nltk.corpus.reader.wordlist.StreamBackedCorpusView.__add__ -nltk.corpus.reader.wordlist.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.wordlist.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.wordlist.StreamBackedCorpusView.__len__ -nltk.corpus.reader.wordlist.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.wordlist.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.wordlist.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.wordlist.StreamBackedCorpusView.fileid -nltk.corpus.reader.wordlist.StreamBackedCorpusView.read_block -nltk.corpus.reader.wordlist.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.wordlist.SyntaxCorpusReader.sents +nltk.corpus.reader.verbnet +nltk.corpus.reader.wordlist nltk.corpus.reader.wordnet.ADJ nltk.corpus.reader.wordnet.ADJ_SAT nltk.corpus.reader.wordnet.ADV -nltk.corpus.reader.wordnet.CorpusReader.citation -nltk.corpus.reader.wordnet.CorpusReader.license -nltk.corpus.reader.wordnet.CorpusReader.raw -nltk.corpus.reader.wordnet.CorpusReader.readme -nltk.corpus.reader.wordnet.CorpusReader.root nltk.corpus.reader.wordnet.Lemma.antonyms nltk.corpus.reader.wordnet.Lemma.count nltk.corpus.reader.wordnet.Lemma.derivationally_related_forms @@ -1497,6 +461,7 @@ nltk.corpus.reader.wordnet.WordNetCorpusReader.synset_from_sense_key nltk.corpus.reader.wordnet.WordNetCorpusReader.words nltk.corpus.reader.wordnet.WordNetCorpusReader.wup_similarity nltk.corpus.reader.wordnet.WordNetError +nltk.corpus.reader.wordnet.WordNetICCorpusReader nltk.corpus.reader.wordnet.information_content nltk.corpus.reader.wordnet.jcn_similarity nltk.corpus.reader.wordnet.lch_similarity @@ -1504,43 +469,8 @@ nltk.corpus.reader.wordnet.lin_similarity nltk.corpus.reader.wordnet.path_similarity nltk.corpus.reader.wordnet.res_similarity nltk.corpus.reader.wordnet.wup_similarity -nltk.corpus.reader.xmldocs.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.xmldocs.ConcatenatedCorpusView.close -nltk.corpus.reader.xmldocs.CorpusReader.citation -nltk.corpus.reader.xmldocs.CorpusReader.license -nltk.corpus.reader.xmldocs.CorpusReader.raw -nltk.corpus.reader.xmldocs.CorpusReader.readme -nltk.corpus.reader.xmldocs.CorpusReader.root -nltk.corpus.reader.xmldocs.StreamBackedCorpusView.__add__ -nltk.corpus.reader.xmldocs.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.xmldocs.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.xmldocs.StreamBackedCorpusView.__len__ -nltk.corpus.reader.xmldocs.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.xmldocs.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.xmldocs.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.xmldocs.StreamBackedCorpusView.fileid -nltk.corpus.reader.xmldocs.StreamBackedCorpusView.read_block -nltk.corpus.reader.ycoe.CategorizedCorpusReader.paras -nltk.corpus.reader.ycoe.CategorizedCorpusReader.raw -nltk.corpus.reader.ycoe.CategorizedCorpusReader.sents -nltk.corpus.reader.ycoe.ConcatenatedCorpusView.__len__ -nltk.corpus.reader.ycoe.ConcatenatedCorpusView.close -nltk.corpus.reader.ycoe.CorpusReader.citation -nltk.corpus.reader.ycoe.CorpusReader.license -nltk.corpus.reader.ycoe.CorpusReader.raw -nltk.corpus.reader.ycoe.CorpusReader.readme -nltk.corpus.reader.ycoe.CorpusReader.root -nltk.corpus.reader.ycoe.StreamBackedCorpusView.__add__ -nltk.corpus.reader.ycoe.StreamBackedCorpusView.__enter__ -nltk.corpus.reader.ycoe.StreamBackedCorpusView.__exit__ -nltk.corpus.reader.ycoe.StreamBackedCorpusView.__len__ -nltk.corpus.reader.ycoe.StreamBackedCorpusView.__mul__ -nltk.corpus.reader.ycoe.StreamBackedCorpusView.__radd__ -nltk.corpus.reader.ycoe.StreamBackedCorpusView.__rmul__ -nltk.corpus.reader.ycoe.StreamBackedCorpusView.fileid -nltk.corpus.reader.ycoe.StreamBackedCorpusView.read_block -nltk.corpus.reader.ycoe.SyntaxCorpusReader.parsed_sents -nltk.corpus.reader.ycoe.SyntaxCorpusReader.sents +nltk.corpus.reader.xmldocs +nltk.corpus.reader.ycoe nltk.corpus.reuters nltk.corpus.rte nltk.corpus.semcor @@ -1575,152 +505,143 @@ nltk.corpus.wordnet2022 nltk.corpus.wordnet31 nltk.corpus.wordnet_ic nltk.corpus.words +nltk.data.AUTO_FORMATS nltk.data.BufferedGzipFile -nltk.downloader.DownloaderGUI.c -nltk.draw.cfg.Nonterminal.__ge__ -nltk.draw.cfg.Nonterminal.__gt__ -nltk.draw.cfg.Nonterminal.__le__ -nltk.featstruct.CustomFeatureValue.__ge__ -nltk.featstruct.CustomFeatureValue.__gt__ -nltk.featstruct.CustomFeatureValue.__le__ -nltk.featstruct.FeatStruct.__ge__ -nltk.featstruct.FeatStruct.__gt__ -nltk.featstruct.FeatStruct.__le__ -nltk.featstruct.Feature.__ge__ -nltk.featstruct.Feature.__gt__ -nltk.featstruct.Feature.__le__ -nltk.grammar.FeatureValueType.__ge__ -nltk.grammar.FeatureValueType.__gt__ -nltk.grammar.FeatureValueType.__le__ -nltk.grammar.Nonterminal.__ge__ -nltk.grammar.Nonterminal.__gt__ -nltk.grammar.Nonterminal.__le__ -nltk.grammar.Production.__ge__ -nltk.grammar.Production.__gt__ -nltk.grammar.Production.__le__ -nltk.inference.nonmonotonic.AbstractVariableExpression.__ge__ -nltk.inference.nonmonotonic.AbstractVariableExpression.__le__ -nltk.inference.nonmonotonic.Variable.__ge__ -nltk.inference.nonmonotonic.Variable.__gt__ -nltk.inference.nonmonotonic.Variable.__le__ -nltk.inference.resolution.Variable.__ge__ -nltk.inference.resolution.Variable.__gt__ -nltk.inference.resolution.Variable.__le__ -nltk.inference.tableau.AbstractVariableExpression.__ge__ -nltk.inference.tableau.AbstractVariableExpression.__le__ -nltk.inference.tableau.Variable.__ge__ -nltk.inference.tableau.Variable.__gt__ -nltk.inference.tableau.Variable.__le__ +nltk.data.FORMATS +nltk.data.FileSystemPathPointer +nltk.data.GzipFileSystemPathPointer +nltk.data.LazyLoader +nltk.data.OpenOnDemandZipFile +nltk.data.__all__ +nltk.data.clear_cache +nltk.data.load +nltk.data.retrieve +nltk.data.show_cfg +nltk.decorators +nltk.demo +nltk.downloader +nltk.draw +nltk.draw.cfg +nltk.draw.dispersion +nltk.draw.table +nltk.draw.tree +nltk.draw.util +nltk.featstruct +nltk.grammar +nltk.help +nltk.inference +nltk.inference.api +nltk.inference.discourse +nltk.inference.mace +nltk.inference.nonmonotonic +nltk.inference.prover9 +nltk.inference.resolution +nltk.inference.tableau nltk.infile -nltk.internals.find_binary -nltk.internals.find_file_iter -nltk.internals.find_jar -nltk.internals.find_jar_iter +nltk.internals +nltk.json_tags +nltk.jsontags nltk.langnames -nltk.log_likelihood -nltk.metrics.BigramAssocMeasures -nltk.metrics.NgramAssocMeasures -nltk.metrics.QuadgramAssocMeasures -nltk.metrics.TrigramAssocMeasures -nltk.metrics.aline.demo -nltk.metrics.association.BigramAssocMeasures -nltk.metrics.association.NgramAssocMeasures -nltk.metrics.association.QuadgramAssocMeasures -nltk.metrics.association.TrigramAssocMeasures -nltk.parse.ProbabilisticNonprojectiveParser.best_incoming_arc -nltk.parse.ProbabilisticNonprojectiveParser.original_best_arc -nltk.parse.chart.EdgeI.__ge__ -nltk.parse.chart.EdgeI.__gt__ -nltk.parse.chart.EdgeI.__le__ -nltk.parse.chart.TreeEdge.move_dot_forward -nltk.parse.corenlp.find_jar_iter -nltk.parse.dependencygraph.find_binary -nltk.parse.earleychart.EdgeI.__ge__ -nltk.parse.earleychart.EdgeI.__gt__ -nltk.parse.earleychart.EdgeI.__le__ -nltk.parse.featurechart.EdgeI.__ge__ -nltk.parse.featurechart.EdgeI.__gt__ -nltk.parse.featurechart.EdgeI.__le__ -nltk.parse.featurechart.FeatStruct.__ge__ -nltk.parse.featurechart.FeatStruct.__gt__ -nltk.parse.featurechart.FeatStruct.__le__ -nltk.parse.featurechart.Nonterminal.__ge__ -nltk.parse.featurechart.Nonterminal.__gt__ -nltk.parse.featurechart.Nonterminal.__le__ -nltk.parse.featurechart.Production.__ge__ -nltk.parse.featurechart.Production.__gt__ -nltk.parse.featurechart.Production.__le__ -nltk.parse.featurechart.TreeEdge.move_dot_forward -nltk.parse.generate.Nonterminal.__ge__ -nltk.parse.generate.Nonterminal.__gt__ -nltk.parse.generate.Nonterminal.__le__ -nltk.parse.nonprojectivedependencyparser.ProbabilisticNonprojectiveParser.best_incoming_arc -nltk.parse.nonprojectivedependencyparser.ProbabilisticNonprojectiveParser.original_best_arc -nltk.parse.pchart.Nonterminal.__ge__ -nltk.parse.pchart.Nonterminal.__gt__ -nltk.parse.pchart.Nonterminal.__le__ -nltk.parse.pchart.TreeEdge.move_dot_forward -nltk.parse.projectivedependencyparser.DependencySpan.__ge__ -nltk.parse.projectivedependencyparser.DependencySpan.__gt__ -nltk.parse.projectivedependencyparser.DependencySpan.__le__ -nltk.parse.recursivedescent.Nonterminal.__ge__ -nltk.parse.recursivedescent.Nonterminal.__gt__ -nltk.parse.recursivedescent.Nonterminal.__le__ -nltk.parse.shiftreduce.Nonterminal.__ge__ -nltk.parse.shiftreduce.Nonterminal.__gt__ -nltk.parse.shiftreduce.Nonterminal.__le__ -nltk.parse.stanford.find_jar_iter -nltk.probability.DictionaryConditionalProbDist -nltk.probability.DictionaryConditionalProbDist.__missing__ -nltk.probability.DictionaryProbDist -nltk.sem.Variable.__ge__ -nltk.sem.Variable.__gt__ -nltk.sem.Variable.__le__ -nltk.sem.boxer.Variable.__ge__ -nltk.sem.boxer.Variable.__gt__ -nltk.sem.boxer.Variable.__le__ -nltk.sem.boxer.find_binary -nltk.sem.cooper_storage.Variable.__ge__ -nltk.sem.cooper_storage.Variable.__gt__ -nltk.sem.cooper_storage.Variable.__le__ -nltk.sem.drt.AbstractVariableExpression.__ge__ -nltk.sem.drt.AbstractVariableExpression.__le__ -nltk.sem.drt.DrsDrawer.__init__ -nltk.sem.drt.Variable.__ge__ -nltk.sem.drt.Variable.__gt__ -nltk.sem.drt.Variable.__le__ -nltk.sem.drt_glue_demo.DrsDrawer.__init__ -nltk.sem.drt_glue_demo.Variable.__ge__ -nltk.sem.drt_glue_demo.Variable.__gt__ -nltk.sem.drt_glue_demo.Variable.__le__ -nltk.sem.evaluate.AbstractVariableExpression.__ge__ -nltk.sem.evaluate.AbstractVariableExpression.__le__ -nltk.sem.evaluate.Variable.__ge__ -nltk.sem.evaluate.Variable.__gt__ -nltk.sem.evaluate.Variable.__le__ -nltk.sem.glue.AbstractVariableExpression.__ge__ -nltk.sem.glue.AbstractVariableExpression.__le__ -nltk.sem.glue.Variable.__ge__ -nltk.sem.glue.Variable.__gt__ -nltk.sem.glue.Variable.__le__ -nltk.sem.logic.AbstractVariableExpression.__ge__ -nltk.sem.logic.AbstractVariableExpression.__le__ -nltk.sem.logic.Variable.__ge__ -nltk.sem.logic.Variable.__gt__ -nltk.sem.logic.Variable.__le__ -nltk.sentiment.sentiment_analyzer.BigramAssocMeasures -nltk.stem.snowball.stopwords -nltk.tag.SennaChunkTagger -nltk.tag.SennaNERTagger -nltk.tag.SennaTagger -nltk.tag.hunpos.find_binary -nltk.tag.senna.Senna -nltk.tag.senna.Senna.SUPPORTED_OPERATIONS -nltk.tag.senna.SennaChunkTagger -nltk.tag.senna.SennaNERTagger -nltk.tag.senna.SennaTagger -nltk.tag.stanford.find_jar -nltk.tbl.demo.treebank +nltk.lazyimport +nltk.lm +nltk.lm.api +nltk.lm.counter +nltk.lm.models +nltk.lm.preprocessing +nltk.lm.smoothing +nltk.lm.util +nltk.lm.vocabulary +nltk.metrics +nltk.metrics.agreement +nltk.metrics.aline +nltk.metrics.association +nltk.metrics.confusionmatrix +nltk.metrics.distance +nltk.metrics.paice +nltk.metrics.scores +nltk.metrics.segmentation +nltk.metrics.spearman +nltk.misc +nltk.misc.babelfish +nltk.misc.chomsky +nltk.misc.minimalset +nltk.misc.sort +nltk.misc.wordfinder +nltk.parse +nltk.parse.api +nltk.parse.bllip +nltk.parse.chart +nltk.parse.corenlp +nltk.parse.dependencygraph +nltk.parse.earleychart +nltk.parse.evaluate +nltk.parse.featurechart +nltk.parse.generate +nltk.parse.malt +nltk.parse.nonprojectivedependencyparser +nltk.parse.pchart +nltk.parse.projectivedependencyparser +nltk.parse.recursivedescent +nltk.parse.shiftreduce +nltk.parse.stanford +nltk.parse.transitionparser +nltk.parse.util +nltk.parse.viterbi +nltk.probability +nltk.sem +nltk.sem.boxer +nltk.sem.chat80 +nltk.sem.cooper_storage +nltk.sem.drt +nltk.sem.drt_glue_demo +nltk.sem.evaluate +nltk.sem.glue +nltk.sem.hole +nltk.sem.lfg +nltk.sem.linearlogic +nltk.sem.logic +nltk.sem.relextract +nltk.sem.skolemize +nltk.sem.util +nltk.sentiment +nltk.sentiment.sentiment_analyzer +nltk.sentiment.util +nltk.sentiment.vader +nltk.stem +nltk.stem.api +nltk.stem.arlstem +nltk.stem.arlstem2 +nltk.stem.cistem +nltk.stem.isri +nltk.stem.lancaster +nltk.stem.porter +nltk.stem.regexp +nltk.stem.rslp +nltk.stem.snowball +nltk.stem.util +nltk.stem.wordnet +nltk.tabdata +nltk.tag +nltk.tag.api +nltk.tag.brill +nltk.tag.brill_trainer +nltk.tag.crf +nltk.tag.hmm +nltk.tag.hunpos +nltk.tag.mapping +nltk.tag.perceptron +nltk.tag.senna +nltk.tag.sequential +nltk.tag.stanford +nltk.tag.tnt +nltk.tag.util +nltk.tbl +nltk.tbl.api +nltk.tbl.demo +nltk.tbl.erroranalysis +nltk.tbl.feature +nltk.tbl.rule +nltk.tbl.template nltk.test nltk.test.all nltk.test.childes_fixt @@ -1785,279 +706,63 @@ nltk.test.unit.translate.test_ibm_model nltk.test.unit.translate.test_meteor nltk.test.unit.translate.test_nist nltk.test.unit.translate.test_stack_decoder -nltk.tokenize.LineTokenizer -nltk.tokenize.LineTokenizer.span_tokenize -nltk.tokenize.LineTokenizer.tokenize -nltk.tokenize.StanfordSegmenter -nltk.tokenize.SyllableTokenizer -nltk.tokenize.TreebankWordTokenizer -nltk.tokenize.TreebankWordTokenizer.CONTRACTIONS2 -nltk.tokenize.TreebankWordTokenizer.CONTRACTIONS3 -nltk.tokenize.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.tokenize.TreebankWordTokenizer.DOUBLE_DASHES -nltk.tokenize.TreebankWordTokenizer.ENDING_QUOTES -nltk.tokenize.TreebankWordTokenizer.PARENS_BRACKETS -nltk.tokenize.TreebankWordTokenizer.PUNCTUATION -nltk.tokenize.TreebankWordTokenizer.STARTING_QUOTES +nltk.text +nltk.tgrep +nltk.tokenize +nltk.tokenize.api +nltk.tokenize.casual +nltk.tokenize.destructive +nltk.tokenize.legality_principle +nltk.tokenize.mwe nltk.tokenize.nist -nltk.tokenize.punkt.PunktLanguageVars.internal_punctuation -nltk.tokenize.punkt.PunktLanguageVars.re_boundary_realignment -nltk.tokenize.punkt.PunktLanguageVars.sent_end_chars -nltk.tokenize.punkt.PunktParameters.clear_abbrevs -nltk.tokenize.punkt.PunktParameters.clear_ortho_context -nltk.tokenize.punkt.PunktToken.abbr -nltk.tokenize.punkt.PunktToken.ellipsis -nltk.tokenize.punkt.PunktToken.is_alpha -nltk.tokenize.punkt.PunktToken.is_non_punct -nltk.tokenize.punkt.PunktToken.linestart -nltk.tokenize.punkt.PunktToken.parastart -nltk.tokenize.punkt.PunktToken.period_final -nltk.tokenize.punkt.PunktToken.sentbreak -nltk.tokenize.punkt.PunktToken.tok -nltk.tokenize.punkt.PunktToken.type -nltk.tokenize.simple.CharTokenizer -nltk.tokenize.simple.LineTokenizer -nltk.tokenize.simple.LineTokenizer.span_tokenize -nltk.tokenize.simple.LineTokenizer.tokenize -nltk.tokenize.sonority_sequencing.SyllableTokenizer -nltk.tokenize.stanford.find_jar -nltk.tokenize.stanford_segmenter.StanfordSegmenter -nltk.tokenize.stanford_segmenter.find_jar -nltk.tokenize.treebank.TreebankWordTokenizer -nltk.tokenize.treebank.TreebankWordTokenizer.CONTRACTIONS2 -nltk.tokenize.treebank.TreebankWordTokenizer.CONTRACTIONS3 -nltk.tokenize.treebank.TreebankWordTokenizer.CONVERT_PARENTHESES -nltk.tokenize.treebank.TreebankWordTokenizer.DOUBLE_DASHES -nltk.tokenize.treebank.TreebankWordTokenizer.ENDING_QUOTES -nltk.tokenize.treebank.TreebankWordTokenizer.PARENS_BRACKETS -nltk.tokenize.treebank.TreebankWordTokenizer.PUNCTUATION -nltk.tokenize.treebank.TreebankWordTokenizer.STARTING_QUOTES -nltk.translate.AlignedSent.alignment -nltk.translate.AlignedSent.invert -nltk.translate.Alignment.__getitem__ -nltk.translate.Alignment.fromstring -nltk.translate.Alignment.invert -nltk.translate.Alignment.range -nltk.translate.IBMModel.MIN_PROB -nltk.translate.IBMModel.maximize_fertility_probabilities -nltk.translate.IBMModel.maximize_lexical_translation_probabilities -nltk.translate.IBMModel.maximize_null_generation_probabilities -nltk.translate.IBMModel.prob_of_alignments -nltk.translate.IBMModel.prob_t_a_given_s -nltk.translate.IBMModel.set_uniform_probabilities -nltk.translate.IBMModel1.prob_alignment_point -nltk.translate.IBMModel1.prob_all_alignments -nltk.translate.IBMModel1.train -nltk.translate.IBMModel2.maximize_alignment_probabilities -nltk.translate.IBMModel2.prob_alignment_point -nltk.translate.IBMModel2.prob_all_alignments -nltk.translate.IBMModel2.train -nltk.translate.IBMModel3.maximize_distortion_probabilities -nltk.translate.IBMModel3.train -nltk.translate.IBMModel4.maximize_distortion_probabilities -nltk.translate.IBMModel4.train -nltk.translate.IBMModel5.MIN_SCORE_FACTOR -nltk.translate.IBMModel5.hillclimb -nltk.translate.IBMModel5.maximize_vacancy_probabilities -nltk.translate.IBMModel5.sample -nltk.translate.IBMModel5.train -nltk.translate.StackDecoder.distortion_factor -nltk.translate.StackDecoder.expansion_score -nltk.translate.StackDecoder.translate -nltk.translate.api.AlignedSent.alignment -nltk.translate.api.AlignedSent.invert -nltk.translate.api.Alignment.__getitem__ -nltk.translate.api.Alignment.fromstring -nltk.translate.api.Alignment.invert -nltk.translate.api.Alignment.range -nltk.translate.api.PhraseTableEntry -nltk.translate.bleu_score.Fraction -nltk.translate.bleu_score.SmoothingFunction.method1 -nltk.translate.bleu_score.SmoothingFunction.method2 -nltk.translate.bleu_score.SmoothingFunction.method3 -nltk.translate.bleu_score.SmoothingFunction.method4 -nltk.translate.bleu_score.SmoothingFunction.method5 -nltk.translate.bleu_score.SmoothingFunction.method6 -nltk.translate.bleu_score.SmoothingFunction.method7 -nltk.translate.gale_church.norm -nltk.translate.ibm1.AlignedSent.alignment -nltk.translate.ibm1.AlignedSent.invert -nltk.translate.ibm1.Alignment.__getitem__ -nltk.translate.ibm1.Alignment.fromstring -nltk.translate.ibm1.Alignment.invert -nltk.translate.ibm1.Alignment.range -nltk.translate.ibm1.IBMModel.MIN_PROB -nltk.translate.ibm1.IBMModel.maximize_fertility_probabilities -nltk.translate.ibm1.IBMModel.maximize_lexical_translation_probabilities -nltk.translate.ibm1.IBMModel.maximize_null_generation_probabilities -nltk.translate.ibm1.IBMModel.prob_of_alignments -nltk.translate.ibm1.IBMModel.prob_t_a_given_s -nltk.translate.ibm1.IBMModel.set_uniform_probabilities -nltk.translate.ibm1.IBMModel1.prob_alignment_point -nltk.translate.ibm1.IBMModel1.prob_all_alignments -nltk.translate.ibm1.IBMModel1.train -nltk.translate.ibm2.AlignedSent.alignment -nltk.translate.ibm2.AlignedSent.invert -nltk.translate.ibm2.Alignment.__getitem__ -nltk.translate.ibm2.Alignment.fromstring -nltk.translate.ibm2.Alignment.invert -nltk.translate.ibm2.Alignment.range -nltk.translate.ibm2.IBMModel.MIN_PROB -nltk.translate.ibm2.IBMModel.maximize_fertility_probabilities -nltk.translate.ibm2.IBMModel.maximize_lexical_translation_probabilities -nltk.translate.ibm2.IBMModel.maximize_null_generation_probabilities -nltk.translate.ibm2.IBMModel.prob_of_alignments -nltk.translate.ibm2.IBMModel.prob_t_a_given_s -nltk.translate.ibm2.IBMModel.set_uniform_probabilities -nltk.translate.ibm2.IBMModel1.prob_alignment_point -nltk.translate.ibm2.IBMModel1.prob_all_alignments -nltk.translate.ibm2.IBMModel1.train -nltk.translate.ibm2.IBMModel2.maximize_alignment_probabilities -nltk.translate.ibm2.IBMModel2.prob_alignment_point -nltk.translate.ibm2.IBMModel2.prob_all_alignments -nltk.translate.ibm2.IBMModel2.train -nltk.translate.ibm3.AlignedSent.alignment -nltk.translate.ibm3.AlignedSent.invert -nltk.translate.ibm3.Alignment.__getitem__ -nltk.translate.ibm3.Alignment.fromstring -nltk.translate.ibm3.Alignment.invert -nltk.translate.ibm3.Alignment.range -nltk.translate.ibm3.IBMModel.MIN_PROB -nltk.translate.ibm3.IBMModel.maximize_fertility_probabilities -nltk.translate.ibm3.IBMModel.maximize_lexical_translation_probabilities -nltk.translate.ibm3.IBMModel.maximize_null_generation_probabilities -nltk.translate.ibm3.IBMModel.prob_of_alignments -nltk.translate.ibm3.IBMModel.prob_t_a_given_s -nltk.translate.ibm3.IBMModel.set_uniform_probabilities -nltk.translate.ibm3.IBMModel2.maximize_alignment_probabilities -nltk.translate.ibm3.IBMModel2.prob_alignment_point -nltk.translate.ibm3.IBMModel2.prob_all_alignments -nltk.translate.ibm3.IBMModel2.train -nltk.translate.ibm3.IBMModel3.maximize_distortion_probabilities -nltk.translate.ibm3.IBMModel3.train -nltk.translate.ibm4.AlignedSent.alignment -nltk.translate.ibm4.AlignedSent.invert -nltk.translate.ibm4.Alignment.__getitem__ -nltk.translate.ibm4.Alignment.fromstring -nltk.translate.ibm4.Alignment.invert -nltk.translate.ibm4.Alignment.range -nltk.translate.ibm4.IBMModel.MIN_PROB -nltk.translate.ibm4.IBMModel.maximize_fertility_probabilities -nltk.translate.ibm4.IBMModel.maximize_lexical_translation_probabilities -nltk.translate.ibm4.IBMModel.maximize_null_generation_probabilities -nltk.translate.ibm4.IBMModel.prob_of_alignments -nltk.translate.ibm4.IBMModel.prob_t_a_given_s -nltk.translate.ibm4.IBMModel.set_uniform_probabilities -nltk.translate.ibm4.IBMModel3.maximize_distortion_probabilities -nltk.translate.ibm4.IBMModel3.train -nltk.translate.ibm4.IBMModel4.maximize_distortion_probabilities -nltk.translate.ibm4.IBMModel4.train -nltk.translate.ibm5.AlignedSent.alignment -nltk.translate.ibm5.AlignedSent.invert -nltk.translate.ibm5.Alignment.__getitem__ -nltk.translate.ibm5.Alignment.fromstring -nltk.translate.ibm5.Alignment.invert -nltk.translate.ibm5.Alignment.range -nltk.translate.ibm5.IBMModel.MIN_PROB -nltk.translate.ibm5.IBMModel.maximize_fertility_probabilities -nltk.translate.ibm5.IBMModel.maximize_lexical_translation_probabilities -nltk.translate.ibm5.IBMModel.maximize_null_generation_probabilities -nltk.translate.ibm5.IBMModel.prob_of_alignments -nltk.translate.ibm5.IBMModel.prob_t_a_given_s -nltk.translate.ibm5.IBMModel.set_uniform_probabilities -nltk.translate.ibm5.IBMModel4.maximize_distortion_probabilities -nltk.translate.ibm5.IBMModel4.train -nltk.translate.ibm5.IBMModel5.MIN_SCORE_FACTOR -nltk.translate.ibm5.IBMModel5.hillclimb -nltk.translate.ibm5.IBMModel5.maximize_vacancy_probabilities -nltk.translate.ibm5.IBMModel5.sample -nltk.translate.ibm5.IBMModel5.train -nltk.translate.ibm_model.AlignmentInfo.zero_indexed_alignment -nltk.translate.ibm_model.IBMModel.MIN_PROB -nltk.translate.ibm_model.IBMModel.maximize_fertility_probabilities -nltk.translate.ibm_model.IBMModel.maximize_lexical_translation_probabilities -nltk.translate.ibm_model.IBMModel.maximize_null_generation_probabilities -nltk.translate.ibm_model.IBMModel.prob_of_alignments -nltk.translate.ibm_model.IBMModel.prob_t_a_given_s -nltk.translate.ibm_model.IBMModel.set_uniform_probabilities -nltk.translate.meteor_score.WordNetCorpusReader.ADJ -nltk.translate.meteor_score.WordNetCorpusReader.ADJ_SAT -nltk.translate.meteor_score.WordNetCorpusReader.ADV -nltk.translate.meteor_score.WordNetCorpusReader.MORPHOLOGICAL_SUBSTITUTIONS -nltk.translate.meteor_score.WordNetCorpusReader.NOUN -nltk.translate.meteor_score.WordNetCorpusReader.VERB -nltk.translate.meteor_score.WordNetCorpusReader.add_exomw -nltk.translate.meteor_score.WordNetCorpusReader.add_omw -nltk.translate.meteor_score.WordNetCorpusReader.add_provs -nltk.translate.meteor_score.WordNetCorpusReader.all_eng_synsets -nltk.translate.meteor_score.WordNetCorpusReader.all_lemma_names -nltk.translate.meteor_score.WordNetCorpusReader.all_omw_synsets -nltk.translate.meteor_score.WordNetCorpusReader.all_synsets -nltk.translate.meteor_score.WordNetCorpusReader.citation -nltk.translate.meteor_score.WordNetCorpusReader.custom_lemmas -nltk.translate.meteor_score.WordNetCorpusReader.digraph -nltk.translate.meteor_score.WordNetCorpusReader.disable_custom_lemmas -nltk.translate.meteor_score.WordNetCorpusReader.doc -nltk.translate.meteor_score.WordNetCorpusReader.ic -nltk.translate.meteor_score.WordNetCorpusReader.jcn_similarity -nltk.translate.meteor_score.WordNetCorpusReader.langs -nltk.translate.meteor_score.WordNetCorpusReader.lch_similarity -nltk.translate.meteor_score.WordNetCorpusReader.lemma -nltk.translate.meteor_score.WordNetCorpusReader.lemma_count -nltk.translate.meteor_score.WordNetCorpusReader.lemma_from_key -nltk.translate.meteor_score.WordNetCorpusReader.lemmas -nltk.translate.meteor_score.WordNetCorpusReader.license -nltk.translate.meteor_score.WordNetCorpusReader.lin_similarity -nltk.translate.meteor_score.WordNetCorpusReader.merged_synsets -nltk.translate.meteor_score.WordNetCorpusReader.morphy -nltk.translate.meteor_score.WordNetCorpusReader.of2ss -nltk.translate.meteor_score.WordNetCorpusReader.path_similarity -nltk.translate.meteor_score.WordNetCorpusReader.readme -nltk.translate.meteor_score.WordNetCorpusReader.res_similarity -nltk.translate.meteor_score.WordNetCorpusReader.split_synsets -nltk.translate.meteor_score.WordNetCorpusReader.ss2of -nltk.translate.meteor_score.WordNetCorpusReader.synonyms -nltk.translate.meteor_score.WordNetCorpusReader.synset -nltk.translate.meteor_score.WordNetCorpusReader.synset_from_sense_key -nltk.translate.meteor_score.WordNetCorpusReader.words -nltk.translate.meteor_score.WordNetCorpusReader.wup_similarity -nltk.translate.meteor_score._enum_align_words -nltk.translate.meteor_score._enum_wordnetsyn_match -nltk.translate.meteor_score.align_words -nltk.translate.meteor_score.meteor_score -nltk.translate.meteor_score.single_meteor_score -nltk.translate.meteor_score.wordnet -nltk.translate.meteor_score.wordnetsyn_match -nltk.translate.ribes_score.spearman_rho -nltk.translate.stack_decoder.StackDecoder.distortion_factor -nltk.translate.stack_decoder.StackDecoder.expansion_score -nltk.translate.stack_decoder.StackDecoder.translate -nltk.translate.stack_decoder._Hypothesis.total_translated_words -nltk.translate.stack_decoder._Stack.__bool__ -nltk.translate.stack_decoder._Stack.__iter__ -nltk.tree.MultiParentedTree -nltk.tree.ParentedTree -nltk.tree.ParentedTree.copy -nltk.tree.parented.AbstractParentedTree -nltk.tree.parented.AbstractParentedTree.__delitem__ -nltk.tree.parented.AbstractParentedTree.__setitem__ -nltk.tree.parented.AbstractParentedTree.append -nltk.tree.parented.AbstractParentedTree.extend -nltk.tree.parented.AbstractParentedTree.insert -nltk.tree.parented.AbstractParentedTree.pop -nltk.tree.parented.AbstractParentedTree.remove -nltk.tree.parented.MultiParentedTree -nltk.tree.parented.ParentedTree -nltk.tree.parented.ParentedTree.copy -nltk.twitter.Authenticate -nltk.twitter.Query -nltk.twitter.Streamer -nltk.twitter.TweetViewer -nltk.twitter.TweetWriter -nltk.twitter.Twitter -nltk.twitter.credsfromfile +nltk.tokenize.punkt +nltk.tokenize.regexp +nltk.tokenize.repp +nltk.tokenize.sexpr +nltk.tokenize.simple +nltk.tokenize.sonority_sequencing +nltk.tokenize.stanford +nltk.tokenize.stanford_segmenter +nltk.tokenize.texttiling +nltk.tokenize.toktok +nltk.tokenize.treebank +nltk.tokenize.util +nltk.toolbox +nltk.translate +nltk.translate.api +nltk.translate.bleu_score +nltk.translate.chrf_score +nltk.translate.gale_church +nltk.translate.gdfa +nltk.translate.gleu_score +nltk.translate.ibm1 +nltk.translate.ibm2 +nltk.translate.ibm3 +nltk.translate.ibm4 +nltk.translate.ibm5 +nltk.translate.ibm_model +nltk.translate.meteor_score +nltk.translate.metrics +nltk.translate.nist_score +nltk.translate.phrase_based +nltk.translate.ribes_score +nltk.translate.stack_decoder +nltk.tree +nltk.tree.immutable +nltk.tree.parented +nltk.tree.parsing +nltk.tree.prettyprinter +nltk.tree.probabilistic +nltk.tree.transforms +nltk.tree.tree +nltk.treeprettyprinter +nltk.treetransforms +nltk.twitter +nltk.twitter.api +nltk.twitter.common nltk.twitter.twitter_demo nltk.twitter.twitterclient nltk.twitter.util +nltk.util nltk.version_file -nltk.wsd.wordnet +nltk.wsd diff --git a/stubs/nltk/app/__init__.pyi b/stubs/nltk/app/__init__.pyi deleted file mode 100644 index 377f60f..0000000 --- a/stubs/nltk/app/__init__.pyi +++ /dev/null @@ -1 +0,0 @@ -from matplotlib import pylab as pylab # type: ignore[import-not-found] diff --git a/stubs/nltk/app/chartparser_app.pyi b/stubs/nltk/app/chartparser_app.pyi deleted file mode 100644 index e4ef55e..0000000 --- a/stubs/nltk/app/chartparser_app.pyi +++ /dev/null @@ -1,169 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.draw import CFGEditor as CFGEditor -from nltk.draw import TreeSegmentWidget as TreeSegmentWidget -from nltk.draw import tree_to_treesegment as tree_to_treesegment -from nltk.draw.util import CanvasFrame as CanvasFrame -from nltk.draw.util import ColorizedList as ColorizedList -from nltk.draw.util import EntryDialog as EntryDialog -from nltk.draw.util import MutableOptionMenu as MutableOptionMenu -from nltk.draw.util import ShowText as ShowText -from nltk.draw.util import SymbolWidget as SymbolWidget -from nltk.grammar import CFG as CFG -from nltk.grammar import Nonterminal as Nonterminal -from nltk.parse.chart import ( - BottomUpPredictCombineRule as BottomUpPredictCombineRule, -) -from nltk.parse.chart import BottomUpPredictRule as BottomUpPredictRule -from nltk.parse.chart import Chart as Chart -from nltk.parse.chart import LeafEdge as LeafEdge -from nltk.parse.chart import LeafInitRule as LeafInitRule -from nltk.parse.chart import ( - SingleEdgeFundamentalRule as SingleEdgeFundamentalRule, -) -from nltk.parse.chart import SteppingChartParser as SteppingChartParser -from nltk.parse.chart import TopDownInitRule as TopDownInitRule -from nltk.parse.chart import TopDownPredictRule as TopDownPredictRule -from nltk.parse.chart import TreeEdge as TreeEdge -from nltk.tree import Tree as Tree -from nltk.util import in_idle as in_idle - -class EdgeList(ColorizedList): - ARROW: Incomplete - -class ChartMatrixView: - def __init__( - self, - parent: Incomplete, - chart: Incomplete, - toplevel: bool = True, - title: str = "Chart Matrix", - show_numedges: bool = False, - ) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def set_chart(self, chart: Incomplete) -> None: ... - def update(self) -> None: ... - def activate(self) -> None: ... - def inactivate(self) -> None: ... - def add_callback(self, event: Incomplete, func: Incomplete) -> None: ... - def remove_callback( - self, event: Incomplete, func: Incomplete | None = None - ) -> None: ... - def select_cell(self, i: Incomplete, j: Incomplete) -> None: ... - def deselect_cell(self) -> None: ... - def view_edge(self, edge: Incomplete) -> None: ... - def mark_edge(self, edge: Incomplete) -> None: ... - def unmark_edge(self, edge: Incomplete | None = None) -> None: ... - def markonly_edge(self, edge: Incomplete) -> None: ... - def draw(self) -> None: ... - def pack(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class ChartResultsView: - def __init__( - self, - parent: Incomplete, - chart: Incomplete, - grammar: Incomplete, - toplevel: bool = True, - ) -> None: ... - def update(self, edge: Incomplete | None = None) -> None: ... - def print_all(self, *e: Incomplete) -> None: ... - def print_selection(self, *e: Incomplete) -> None: ... - def clear(self) -> None: ... - def set_chart(self, chart: Incomplete) -> None: ... - def set_grammar(self, grammar: Incomplete) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def pack(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class ChartComparer: - def __init__(self, *chart_filename: Incomplete) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - CHART_FILE_TYPES: Incomplete - def save_chart_dialog(self, *args: Incomplete) -> None: ... - def load_chart_dialog(self, *args: Incomplete) -> None: ... - def load_chart(self, filename: Incomplete) -> None: ... - def select_edge(self, edge: Incomplete) -> None: ... - def select_cell(self, i: Incomplete, j: Incomplete) -> None: ... - -class ChartView: - def __init__( - self, - chart: Incomplete, - root: Incomplete | None = None, - **kw: Incomplete, - ) -> None: ... - def scroll_up(self, *e: Incomplete) -> None: ... - def scroll_down(self, *e: Incomplete) -> None: ... - def page_up(self, *e: Incomplete) -> None: ... - def page_down(self, *e: Incomplete) -> None: ... - def set_font_size(self, size: Incomplete) -> None: ... - def get_font_size(self) -> Incomplete: ... - def update(self, chart: Incomplete | None = None) -> None: ... - def view_edge(self, edge: Incomplete) -> None: ... - def mark_edge(self, edge: Incomplete, mark: str = "#0df") -> None: ... - def unmark_edge(self, edge: Incomplete | None = None) -> None: ... - def markonly_edge(self, edge: Incomplete, mark: str = "#0df") -> None: ... - def erase_tree(self) -> None: ... - def draw_tree(self, edge: Incomplete | None = None) -> None: ... - def cycle_tree(self) -> None: ... - def draw(self) -> None: ... - def add_callback(self, event: Incomplete, func: Incomplete) -> None: ... - def remove_callback( - self, event: Incomplete, func: Incomplete | None = None - ) -> None: ... - -class EdgeRule: - NUM_EDGES: Incomplete - def __init__(self, edge: Incomplete) -> None: ... - def apply( - self, chart: Incomplete, grammar: Incomplete, *e: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... - -class TopDownPredictEdgeRule(EdgeRule, TopDownPredictRule): ... # type: ignore[misc] -class BottomUpEdgeRule(EdgeRule, BottomUpPredictRule): ... # type: ignore[misc] -class BottomUpLeftCornerEdgeRule(EdgeRule, BottomUpPredictCombineRule): ... # type: ignore[misc] -class FundamentalEdgeRule(EdgeRule, SingleEdgeFundamentalRule): ... # type: ignore[misc] - -class ChartParserApp: - def __init__( - self, - grammar: Incomplete, - tokens: Incomplete, - title: str = "Chart Parser Application", - ) -> None: ... - def destroy(self, *args: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def help(self, *e: Incomplete) -> None: ... - def about(self, *e: Incomplete) -> None: ... - CHART_FILE_TYPES: Incomplete - GRAMMAR_FILE_TYPES: Incomplete - def load_chart(self, *args: Incomplete) -> None: ... - def save_chart(self, *args: Incomplete) -> None: ... - def load_grammar(self, *args: Incomplete) -> None: ... - def save_grammar(self, *args: Incomplete) -> None: ... - def reset(self, *args: Incomplete) -> None: ... - def edit_grammar(self, *e: Incomplete) -> None: ... - def set_grammar(self, grammar: Incomplete) -> None: ... - def edit_sentence(self, *e: Incomplete) -> None: ... - def set_sentence(self, sentence: Incomplete) -> None: ... - def view_matrix(self, *e: Incomplete) -> None: ... - def view_results(self, *e: Incomplete) -> None: ... - def resize(self) -> None: ... - def set_font_size(self, size: Incomplete) -> None: ... - def get_font_size(self) -> Incomplete: ... - def apply_strategy( - self, strategy: Incomplete, edge_strategy: Incomplete | None = None - ) -> None: ... - def top_down_init(self, *e: Incomplete) -> None: ... - def top_down_predict(self, *e: Incomplete) -> None: ... - def bottom_up(self, *e: Incomplete) -> None: ... - def bottom_up_leftcorner(self, *e: Incomplete) -> None: ... - def fundamental(self, *e: Incomplete) -> None: ... - def bottom_up_strategy(self, *e: Incomplete) -> None: ... - def bottom_up_leftcorner_strategy(self, *e: Incomplete) -> None: ... - def top_down_strategy(self, *e: Incomplete) -> None: ... - -def app() -> None: ... diff --git a/stubs/nltk/app/chunkparser_app.pyi b/stubs/nltk/app/chunkparser_app.pyi deleted file mode 100644 index 0674e42..0000000 --- a/stubs/nltk/app/chunkparser_app.pyi +++ /dev/null @@ -1,52 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chunk import ChunkScore as ChunkScore -from nltk.chunk import RegexpChunkParser as RegexpChunkParser -from nltk.chunk.regexp import RegexpChunkRule as RegexpChunkRule -from nltk.corpus import conll2000 as conll2000 -from nltk.corpus import treebank_chunk as treebank_chunk -from nltk.draw.util import ShowText as ShowText -from nltk.tree import Tree as Tree -from nltk.util import in_idle as in_idle - -class RegexpChunkApp: - TAGSET: Incomplete - HELP: Incomplete - HELP_AUTOTAG: Incomplete - def normalize_grammar(self, grammar: Incomplete) -> Incomplete: ... - tagset: Incomplete - chunker: Incomplete - grammar: Incomplete - normalized_grammar: Incomplete - grammar_changed: int - devset: Incomplete - devset_name: Incomplete - devset_index: int - def __init__( - self, - devset_name: str = "conll2000", - devset: Incomplete | None = None, - grammar: str = "", - chunk_label: str = "NP", - tagset: Incomplete | None = None, - ) -> None: ... - def toggle_show_trace(self, *e: Incomplete) -> Incomplete: ... - charnum: Incomplete - linenum: Incomplete - def show_trace(self, *e: Incomplete) -> None: ... - def show_help(self, tab: Incomplete) -> Incomplete: ... - top: Incomplete - def destroy(self, *e: Incomplete) -> None: ... - def show_devset(self, index: Incomplete | None = None) -> None: ... - def update(self, *event: Incomplete) -> None: ... - def reset(self) -> None: ... - SAVE_GRAMMAR_TEMPLATE: str - def save_grammar(self, filename: Incomplete | None = None) -> None: ... - def load_grammar(self, filename: Incomplete | None = None) -> None: ... - def save_history(self, filename: Incomplete | None = None) -> None: ... - def about(self, *e: Incomplete) -> None: ... - def set_devset_size(self, size: Incomplete | None = None) -> None: ... - def resize(self, size: Incomplete | None = None) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -def app() -> None: ... diff --git a/stubs/nltk/app/collocations_app.pyi b/stubs/nltk/app/collocations_app.pyi deleted file mode 100644 index 1fb0759..0000000 --- a/stubs/nltk/app/collocations_app.pyi +++ /dev/null @@ -1,67 +0,0 @@ -import threading - -from _typeshed import Incomplete - -from nltk.corpus import alpino as alpino -from nltk.corpus import brown as brown -from nltk.corpus import cess_cat as cess_cat -from nltk.corpus import cess_esp as cess_esp -from nltk.corpus import floresta as floresta -from nltk.corpus import indian as indian -from nltk.corpus import mac_morpho as mac_morpho -from nltk.corpus import machado as machado -from nltk.corpus import nps_chat as nps_chat -from nltk.corpus import sinica_treebank as sinica_treebank -from nltk.corpus import treebank as treebank -from nltk.probability import FreqDist as FreqDist -from nltk.util import in_idle as in_idle - -CORPUS_LOADED_EVENT: str -ERROR_LOADING_CORPUS_EVENT: str -POLL_INTERVAL: int - -class CollocationsView: - queue: Incomplete - model: Incomplete - top: Incomplete - after: Incomplete - def __init__(self) -> None: ... - def set_result_size(self, **kwargs: Incomplete) -> None: ... - current_page: int - def reset_current_page(self) -> None: ... - def handle_error_loading_corpus(self, event: Incomplete) -> None: ... - def handle_corpus_loaded(self, event: Incomplete) -> None: ... - def corpus_selected(self, *args: Incomplete) -> None: ... - def previous(self) -> None: ... - def __next__(self) -> None: ... - def load_corpus(self, selection: Incomplete) -> None: ... - def freeze_editable(self) -> None: ... - def clear_results_box(self) -> None: ... - def fire_event(self, event: Incomplete) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def unfreeze_editable(self) -> None: ... - def set_paging_button_states(self) -> None: ... - def write_results(self, results: Incomplete) -> None: ... - -class CollocationsModel: - result_count: Incomplete - selected_corpus: Incomplete - collocations: Incomplete - CORPORA: Incomplete - DEFAULT_CORPUS: Incomplete - queue: Incomplete - def __init__(self, queue: Incomplete) -> None: ... - result_pages: Incomplete - results_returned: int - def reset_results(self) -> None: ... - def load_corpus(self, name: Incomplete) -> None: ... - def non_default_corpora(self) -> Incomplete: ... - def is_last_page(self, number: Incomplete) -> Incomplete: ... - def next(self, page: Incomplete) -> Incomplete: ... - def prev(self, page: Incomplete) -> Incomplete: ... - class LoadCorpus(threading.Thread): - def __init__(self, name: Incomplete, model: Incomplete) -> None: ... - def run(self) -> None: ... - -def app() -> None: ... diff --git a/stubs/nltk/app/concordance_app.pyi b/stubs/nltk/app/concordance_app.pyi deleted file mode 100644 index dbaeae9..0000000 --- a/stubs/nltk/app/concordance_app.pyi +++ /dev/null @@ -1,97 +0,0 @@ -import threading - -from _typeshed import Incomplete - -from nltk.corpus import alpino as alpino -from nltk.corpus import brown as brown -from nltk.corpus import cess_cat as cess_cat -from nltk.corpus import cess_esp as cess_esp -from nltk.corpus import floresta as floresta -from nltk.corpus import indian as indian -from nltk.corpus import mac_morpho as mac_morpho -from nltk.corpus import nps_chat as nps_chat -from nltk.corpus import sinica_treebank as sinica_treebank -from nltk.corpus import treebank as treebank -from nltk.draw.util import ShowText as ShowText -from nltk.util import in_idle as in_idle - -WORD_OR_TAG: str -BOUNDARY: str -CORPUS_LOADED_EVENT: str -SEARCH_TERMINATED_EVENT: str -SEARCH_ERROR_EVENT: str -ERROR_LOADING_CORPUS_EVENT: str -POLL_INTERVAL: int - -class ConcordanceSearchView: - queue: Incomplete - model: Incomplete - top: Incomplete - after: Incomplete - def __init__(self) -> None: ... - def set_result_size(self, **kwargs: Incomplete) -> None: ... - def set_cntx_af_len(self, **kwargs: Incomplete) -> None: ... - def set_cntx_bf_len(self, **kwargs: Incomplete) -> None: ... - def search_enter_keypress_handler(self, *event: Incomplete) -> None: ... - def previous(self) -> None: ... - def __next__(self) -> None: ... - def about(self, *e: Incomplete) -> None: ... - def handle_error_loading_corpus(self, event: Incomplete) -> None: ... - def handle_corpus_loaded(self, event: Incomplete) -> None: ... - current_page: Incomplete - def handle_search_terminated(self, event: Incomplete) -> None: ... - def handle_search_error(self, event: Incomplete) -> None: ... - def corpus_selected(self, *args: Incomplete) -> None: ... - def load_corpus(self, selection: Incomplete) -> None: ... - def search(self) -> None: ... - def write_results(self, results: Incomplete) -> None: ... - def words_and_labels( - self, sentence: Incomplete, pos1: Incomplete, pos2: Incomplete - ) -> Incomplete: ... - def pad( - self, sent: Incomplete, hstart: Incomplete, hend: Incomplete - ) -> Incomplete: ... - def destroy(self, *e: Incomplete) -> None: ... - def clear_all(self) -> None: ... - def clear_results_box(self) -> None: ... - def freeze_editable(self) -> None: ... - def unfreeze_editable(self) -> None: ... - def set_paging_button_states(self) -> None: ... - def fire_event(self, event: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class ConcordanceSearchModel: - queue: Incomplete - CORPORA: Incomplete - DEFAULT_CORPUS: Incomplete - selected_corpus: Incomplete - result_count: Incomplete - last_sent_searched: int - def __init__(self, queue: Incomplete) -> None: ... - def non_default_corpora(self) -> Incomplete: ... - tagged_sents: Incomplete - def load_corpus(self, name: Incomplete) -> None: ... - query: Incomplete - last_requested_page: Incomplete - def search(self, query: Incomplete, page: Incomplete) -> None: ... - def next(self, page: Incomplete) -> None: ... - def prev(self, page: Incomplete) -> None: ... - results: Incomplete - last_page: Incomplete - def reset_results(self) -> None: ... - def reset_query(self) -> None: ... - def set_results(self, page: Incomplete, resultset: Incomplete) -> None: ... - def get_results(self) -> Incomplete: ... - def has_more_pages(self, page: Incomplete) -> Incomplete: ... - class LoadCorpus(threading.Thread): - def __init__(self, name: Incomplete, model: Incomplete) -> None: ... - def run(self) -> None: ... - - class SearchCorpus(threading.Thread): - def __init__( - self, model: Incomplete, page: Incomplete, count: Incomplete - ) -> None: ... - def run(self) -> None: ... - def processed_query(self) -> Incomplete: ... - -def app() -> None: ... diff --git a/stubs/nltk/app/nemo_app.pyi b/stubs/nltk/app/nemo_app.pyi deleted file mode 100644 index 2eccf57..0000000 --- a/stubs/nltk/app/nemo_app.pyi +++ /dev/null @@ -1,44 +0,0 @@ -from _typeshed import Incomplete - -windowTitle: str -initialFind: str -initialRepl: str -initialText: str -images: Incomplete -colors: Incomplete -emphColors: Incomplete -fieldParams: Incomplete -textParams: Incomplete - -class Zone: - image: Incomplete - imageDimmed: Incomplete - img: Incomplete - fld: Incomplete - txt: Incomplete - def __init__( - self, - image: Incomplete, - initialField: Incomplete, - initialText: Incomplete, - ) -> None: ... - def initScrollText( - self, frm: Incomplete, txt: Incomplete, contents: Incomplete - ) -> None: ... - colorCycle: Incomplete - def refresh(self) -> None: ... - -class FindZone(Zone): - def addTags(self, m: Incomplete) -> None: ... - rex: Incomplete - rexSel: Incomplete - def substitute(self, *args: Incomplete) -> None: ... - -class ReplaceZone(Zone): - def addTags(self, m: Incomplete) -> None: ... - diff: int - repl: Incomplete - def substitute(self) -> None: ... - -def launchRefresh(_: Incomplete) -> None: ... -def app() -> None: ... diff --git a/stubs/nltk/app/rdparser_app.pyi b/stubs/nltk/app/rdparser_app.pyi deleted file mode 100644 index 2fc79b5..0000000 --- a/stubs/nltk/app/rdparser_app.pyi +++ /dev/null @@ -1,38 +0,0 @@ -from _typeshed import Incomplete - -from nltk.draw import CFGEditor as CFGEditor -from nltk.draw import TreeSegmentWidget as TreeSegmentWidget -from nltk.draw import tree_to_treesegment as tree_to_treesegment -from nltk.draw.util import CanvasFrame as CanvasFrame -from nltk.draw.util import EntryDialog as EntryDialog -from nltk.draw.util import ShowText as ShowText -from nltk.draw.util import TextWidget as TextWidget -from nltk.parse import ( - SteppingRecursiveDescentParser as SteppingRecursiveDescentParser, -) -from nltk.tree import Tree as Tree -from nltk.util import in_idle as in_idle - -class RecursiveDescentApp: - def __init__( - self, grammar: Incomplete, sent: Incomplete, trace: int = 0 - ) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def reset(self, *e: Incomplete) -> None: ... - def autostep(self, *e: Incomplete) -> None: ... - def cancel_autostep(self, *e: Incomplete) -> None: ... - def step(self, *e: Incomplete) -> None: ... - def match(self, *e: Incomplete) -> None: ... - def expand(self, *e: Incomplete) -> None: ... - def backtrack(self, *e: Incomplete) -> None: ... - def about(self, *e: Incomplete) -> None: ... - def help(self, *e: Incomplete) -> None: ... - def postscript(self, *e: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def resize(self, size: Incomplete | None = None) -> None: ... - def edit_grammar(self, *e: Incomplete) -> None: ... - def set_grammar(self, grammar: Incomplete) -> None: ... - def edit_sentence(self, *e: Incomplete) -> None: ... - def set_sentence(self, sentence: Incomplete) -> None: ... - -def app() -> None: ... diff --git a/stubs/nltk/app/srparser_app.pyi b/stubs/nltk/app/srparser_app.pyi deleted file mode 100644 index 0c9fb8f..0000000 --- a/stubs/nltk/app/srparser_app.pyi +++ /dev/null @@ -1,34 +0,0 @@ -from _typeshed import Incomplete - -from nltk.draw import CFGEditor as CFGEditor -from nltk.draw import TreeSegmentWidget as TreeSegmentWidget -from nltk.draw import tree_to_treesegment as tree_to_treesegment -from nltk.draw.util import CanvasFrame as CanvasFrame -from nltk.draw.util import EntryDialog as EntryDialog -from nltk.draw.util import ShowText as ShowText -from nltk.draw.util import TextWidget as TextWidget -from nltk.parse import SteppingShiftReduceParser as SteppingShiftReduceParser -from nltk.tree import Tree as Tree -from nltk.util import in_idle as in_idle - -class ShiftReduceApp: - def __init__( - self, grammar: Incomplete, sent: Incomplete, trace: int = 0 - ) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def reset(self, *e: Incomplete) -> None: ... - def step(self, *e: Incomplete) -> Incomplete: ... - def shift(self, *e: Incomplete) -> Incomplete: ... - def reduce(self, *e: Incomplete) -> Incomplete: ... - def undo(self, *e: Incomplete) -> None: ... - def postscript(self, *e: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def resize(self, size: Incomplete | None = None) -> None: ... - def help(self, *e: Incomplete) -> None: ... - def about(self, *e: Incomplete) -> None: ... - def edit_grammar(self, *e: Incomplete) -> None: ... - def set_grammar(self, grammar: Incomplete) -> None: ... - def edit_sentence(self, *e: Incomplete) -> None: ... - def set_sentence(self, sent: Incomplete) -> None: ... - -def app() -> None: ... diff --git a/stubs/nltk/app/wordfreq_app.pyi b/stubs/nltk/app/wordfreq_app.pyi deleted file mode 100644 index 000f78b..0000000 --- a/stubs/nltk/app/wordfreq_app.pyi +++ /dev/null @@ -1,3 +0,0 @@ -__all__ = ["app"] - -def app() -> None: ... diff --git a/stubs/nltk/app/wordnet_app.pyi b/stubs/nltk/app/wordnet_app.pyi deleted file mode 100644 index a08f66a..0000000 --- a/stubs/nltk/app/wordnet_app.pyi +++ /dev/null @@ -1,31 +0,0 @@ -import pickle -from http.server import BaseHTTPRequestHandler - -from _typeshed import Incomplete - -__all__ = ["app"] - -class MyServerHandler(BaseHTTPRequestHandler): - def do_HEAD(self) -> None: ... - def do_GET(self) -> None: ... - def send_head(self, type: Incomplete | None = None) -> None: ... - def log_message(self, format: Incomplete, *args: Incomplete) -> None: ... - -class RestrictedUnpickler(pickle.Unpickler): - def find_class(self, module: Incomplete, name: Incomplete) -> None: ... - -class Reference: - word: Incomplete - synset_relations: Incomplete - def __init__( - self, word: Incomplete, synset_relations: Incomplete = ... - ) -> None: ... - def encode(self) -> Incomplete: ... - @staticmethod - def decode(string: Incomplete) -> Incomplete: ... - def toggle_synset_relation( - self, synset: Incomplete, relation: Incomplete - ) -> Incomplete: ... - def toggle_synset(self, synset: Incomplete) -> Incomplete: ... - -def app() -> None: ... diff --git a/stubs/nltk/book.pyi b/stubs/nltk/book.pyi deleted file mode 100644 index 0a8c5e0..0000000 --- a/stubs/nltk/book.pyi +++ /dev/null @@ -1,36 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus import genesis as genesis -from nltk.corpus import gutenberg as gutenberg -from nltk.corpus import inaugural as inaugural -from nltk.corpus import nps_chat as nps_chat -from nltk.corpus import treebank as treebank -from nltk.corpus import webtext as webtext -from nltk.corpus import wordnet as wordnet -from nltk.probability import FreqDist as FreqDist -from nltk.text import Text as Text -from nltk.util import bigrams as bigrams - -text1: Incomplete -text2: Incomplete -text3: Incomplete -text4: Incomplete -text5: Incomplete -text6: Incomplete -text7: Incomplete -text8: Incomplete -text9: Incomplete - -def texts() -> None: ... - -sent1: Incomplete -sent2: Incomplete -sent3: Incomplete -sent4: Incomplete -sent5: Incomplete -sent6: Incomplete -sent7: Incomplete -sent8: Incomplete -sent9: Incomplete - -def sents() -> None: ... diff --git a/stubs/nltk/ccg/__init__.pyi b/stubs/nltk/ccg/__init__.pyi deleted file mode 100644 index 067783c..0000000 --- a/stubs/nltk/ccg/__init__.pyi +++ /dev/null @@ -1,30 +0,0 @@ -from nltk.ccg.chart import CCGChart as CCGChart -from nltk.ccg.chart import CCGChartParser as CCGChartParser -from nltk.ccg.chart import CCGEdge as CCGEdge -from nltk.ccg.chart import CCGLeafEdge as CCGLeafEdge -from nltk.ccg.combinator import BackwardApplication as BackwardApplication -from nltk.ccg.combinator import BackwardBx as BackwardBx -from nltk.ccg.combinator import BackwardCombinator as BackwardCombinator -from nltk.ccg.combinator import BackwardComposition as BackwardComposition -from nltk.ccg.combinator import BackwardSx as BackwardSx -from nltk.ccg.combinator import BackwardT as BackwardT -from nltk.ccg.combinator import ( - DirectedBinaryCombinator as DirectedBinaryCombinator, -) -from nltk.ccg.combinator import ForwardApplication as ForwardApplication -from nltk.ccg.combinator import ForwardCombinator as ForwardCombinator -from nltk.ccg.combinator import ForwardComposition as ForwardComposition -from nltk.ccg.combinator import ForwardSubstitution as ForwardSubstitution -from nltk.ccg.combinator import ForwardT as ForwardT -from nltk.ccg.combinator import ( - UndirectedBinaryCombinator as UndirectedBinaryCombinator, -) -from nltk.ccg.combinator import UndirectedComposition as UndirectedComposition -from nltk.ccg.combinator import ( - UndirectedFunctionApplication as UndirectedFunctionApplication, -) -from nltk.ccg.combinator import ( - UndirectedSubstitution as UndirectedSubstitution, -) -from nltk.ccg.combinator import UndirectedTypeRaise as UndirectedTypeRaise -from nltk.ccg.lexicon import CCGLexicon as CCGLexicon diff --git a/stubs/nltk/ccg/api.pyi b/stubs/nltk/ccg/api.pyi deleted file mode 100644 index 4845032..0000000 --- a/stubs/nltk/ccg/api.pyi +++ /dev/null @@ -1,76 +0,0 @@ -from abc import ABCMeta, abstractmethod - -from _typeshed import Incomplete - -from nltk.internals import raise_unorderable_types as raise_unorderable_types - -class AbstractCCGCategory(metaclass=ABCMeta): - @abstractmethod - def is_primitive(self) -> Incomplete: ... - @abstractmethod - def is_function(self) -> Incomplete: ... - @abstractmethod - def is_var(self) -> Incomplete: ... - @abstractmethod - def substitute(self, substitutions: Incomplete) -> Incomplete: ... - @abstractmethod - def can_unify(self, other: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class CCGVar(AbstractCCGCategory): - def __init__(self, prim_only: bool = False) -> None: ... - @classmethod - def new_id(cls: Incomplete) -> Incomplete: ... - @classmethod - def reset_id(cls: Incomplete) -> None: ... - def is_primitive(self) -> Incomplete: ... - def is_function(self) -> Incomplete: ... - def is_var(self) -> Incomplete: ... - def substitute(self, substitutions: Incomplete) -> Incomplete: ... - def can_unify(self, other: Incomplete) -> Incomplete: ... - def id(self) -> Incomplete: ... - -class Direction: - def __init__(self, dir: Incomplete, restrictions: Incomplete) -> None: ... - def is_forward(self) -> Incomplete: ... - def is_backward(self) -> Incomplete: ... - def dir(self) -> Incomplete: ... - def restrs(self) -> Incomplete: ... - def is_variable(self) -> Incomplete: ... - def can_unify(self, other: Incomplete) -> Incomplete: ... - def substitute(self, subs: Incomplete) -> Incomplete: ... - def can_compose(self) -> Incomplete: ... - def can_cross(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - def __neg__(self) -> Incomplete: ... - -class PrimitiveCategory(AbstractCCGCategory): - def __init__( - self, categ: Incomplete, restrictions: Incomplete = [] - ) -> None: ... - def is_primitive(self) -> Incomplete: ... - def is_function(self) -> Incomplete: ... - def is_var(self) -> Incomplete: ... - def restrs(self) -> Incomplete: ... - def categ(self) -> Incomplete: ... - def substitute(self, subs: Incomplete) -> Incomplete: ... - def can_unify(self, other: Incomplete) -> Incomplete: ... - -class FunctionalCategory(AbstractCCGCategory): - def __init__( - self, res: Incomplete, arg: Incomplete, dir: Incomplete - ) -> None: ... - def is_primitive(self) -> Incomplete: ... - def is_function(self) -> Incomplete: ... - def is_var(self) -> Incomplete: ... - def substitute(self, subs: Incomplete) -> Incomplete: ... - def can_unify(self, other: Incomplete) -> Incomplete: ... - def arg(self) -> Incomplete: ... - def res(self) -> Incomplete: ... - def dir(self) -> Incomplete: ... diff --git a/stubs/nltk/ccg/chart.pyi b/stubs/nltk/ccg/chart.pyi deleted file mode 100644 index 2c39d06..0000000 --- a/stubs/nltk/ccg/chart.pyi +++ /dev/null @@ -1,117 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.ccg.combinator import * -from nltk.ccg.combinator import BackwardApplication as BackwardApplication -from nltk.ccg.combinator import BackwardBx as BackwardBx -from nltk.ccg.combinator import BackwardComposition as BackwardComposition -from nltk.ccg.combinator import BackwardSx as BackwardSx -from nltk.ccg.combinator import BackwardT as BackwardT -from nltk.ccg.combinator import ForwardApplication as ForwardApplication -from nltk.ccg.combinator import ForwardComposition as ForwardComposition -from nltk.ccg.combinator import ForwardSubstitution as ForwardSubstitution -from nltk.ccg.combinator import ForwardT as ForwardT -from nltk.ccg.lexicon import Token as Token -from nltk.ccg.lexicon import fromstring as fromstring -from nltk.ccg.logic import * -from nltk.parse import ParserI as ParserI -from nltk.parse.chart import AbstractChartRule as AbstractChartRule -from nltk.parse.chart import Chart as Chart -from nltk.parse.chart import EdgeI as EdgeI -from nltk.sem.logic import * -from nltk.tree import Tree as Tree - -class CCGEdge(EdgeI): - def __init__( - self, span: Incomplete, categ: Incomplete, rule: Incomplete - ) -> None: ... - def lhs(self) -> Incomplete: ... - def span(self) -> Incomplete: ... - def start(self) -> Incomplete: ... - def end(self) -> Incomplete: ... - def length(self) -> Incomplete: ... - def rhs(self) -> Incomplete: ... - def dot(self) -> Incomplete: ... - def is_complete(self) -> Incomplete: ... - def is_incomplete(self) -> Incomplete: ... - def nextsym(self) -> None: ... - def categ(self) -> Incomplete: ... - def rule(self) -> Incomplete: ... - -class CCGLeafEdge(EdgeI): - def __init__( - self, pos: Incomplete, token: Incomplete, leaf: Incomplete - ) -> None: ... - def lhs(self) -> Incomplete: ... - def span(self) -> Incomplete: ... - def start(self) -> Incomplete: ... - def end(self) -> Incomplete: ... - def length(self) -> Incomplete: ... - def rhs(self) -> Incomplete: ... - def dot(self) -> Incomplete: ... - def is_complete(self) -> Incomplete: ... - def is_incomplete(self) -> Incomplete: ... - def nextsym(self) -> None: ... - def token(self) -> Incomplete: ... - def categ(self) -> Incomplete: ... - def leaf(self) -> Incomplete: ... - -class BinaryCombinatorRule(AbstractChartRule): - NUMEDGES: int - def __init__(self, combinator: Incomplete) -> None: ... - def apply( - self, - chart: Incomplete, - grammar: Incomplete, - left_edge: Incomplete, - right_edge: Incomplete, - ) -> Generator[Incomplete, None, None]: ... - -class ForwardTypeRaiseRule(AbstractChartRule): - NUMEDGES: int - def __init__(self) -> None: ... - def apply( - self, - chart: Incomplete, - grammar: Incomplete, - left_edge: Incomplete, - right_edge: Incomplete, - ) -> Generator[Incomplete, None, None]: ... - -class BackwardTypeRaiseRule(AbstractChartRule): - NUMEDGES: int - def __init__(self) -> None: ... - def apply( - self, - chart: Incomplete, - grammar: Incomplete, - left_edge: Incomplete, - right_edge: Incomplete, - ) -> Generator[Incomplete, None, None]: ... - -ApplicationRuleSet: Incomplete -CompositionRuleSet: Incomplete -SubstitutionRuleSet: Incomplete -TypeRaiseRuleSet: Incomplete -DefaultRuleSet: Incomplete - -class CCGChartParser(ParserI): - def __init__( - self, lexicon: Incomplete, rules: Incomplete, trace: int = 0 - ) -> None: ... - def lexicon(self) -> Incomplete: ... - def parse(self, tokens: Incomplete) -> Incomplete: ... - -class CCGChart(Chart): - def __init__(self, tokens: Incomplete) -> None: ... - -def compute_semantics( - children: Incomplete, edge: Incomplete -) -> Incomplete: ... -def printCCGDerivation(tree: Incomplete) -> None: ... -def printCCGTree(lwidth: Incomplete, tree: Incomplete) -> Incomplete: ... - -lex: Incomplete - -def demo() -> None: ... diff --git a/stubs/nltk/ccg/combinator.pyi b/stubs/nltk/ccg/combinator.pyi deleted file mode 100644 index cd5b23f..0000000 --- a/stubs/nltk/ccg/combinator.pyi +++ /dev/null @@ -1,111 +0,0 @@ -from abc import ABCMeta, abstractmethod -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.ccg.api import FunctionalCategory as FunctionalCategory - -class UndirectedBinaryCombinator(metaclass=ABCMeta): - @abstractmethod - def can_combine( - self, function: Incomplete, argument: Incomplete - ) -> Incomplete: ... - @abstractmethod - def combine( - self, function: Incomplete, argument: Incomplete - ) -> Incomplete: ... - -class DirectedBinaryCombinator(metaclass=ABCMeta): - @abstractmethod - def can_combine( - self, left: Incomplete, right: Incomplete - ) -> Incomplete: ... - @abstractmethod - def combine(self, left: Incomplete, right: Incomplete) -> Incomplete: ... - -class ForwardCombinator(DirectedBinaryCombinator): - def __init__( - self, combinator: Incomplete, predicate: Incomplete, suffix: str = "" - ) -> None: ... - def can_combine( - self, left: Incomplete, right: Incomplete - ) -> Incomplete: ... - def combine( - self, left: Incomplete, right: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... - -class BackwardCombinator(DirectedBinaryCombinator): - def __init__( - self, combinator: Incomplete, predicate: Incomplete, suffix: str = "" - ) -> None: ... - def can_combine( - self, left: Incomplete, right: Incomplete - ) -> Incomplete: ... - def combine( - self, left: Incomplete, right: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... - -class UndirectedFunctionApplication(UndirectedBinaryCombinator): - def can_combine( - self, function: Incomplete, argument: Incomplete - ) -> Incomplete: ... - def combine( - self, function: Incomplete, argument: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -def forwardOnly(left: Incomplete, right: Incomplete) -> Incomplete: ... -def backwardOnly(left: Incomplete, right: Incomplete) -> Incomplete: ... - -ForwardApplication: Incomplete -BackwardApplication: Incomplete - -class UndirectedComposition(UndirectedBinaryCombinator): - def can_combine( - self, function: Incomplete, argument: Incomplete - ) -> Incomplete: ... - def combine( - self, function: Incomplete, argument: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -def bothForward(left: Incomplete, right: Incomplete) -> Incomplete: ... -def bothBackward(left: Incomplete, right: Incomplete) -> Incomplete: ... -def crossedDirs(left: Incomplete, right: Incomplete) -> Incomplete: ... -def backwardBxConstraint( - left: Incomplete, right: Incomplete -) -> Incomplete: ... - -ForwardComposition: Incomplete -BackwardComposition: Incomplete -BackwardBx: Incomplete - -class UndirectedSubstitution(UndirectedBinaryCombinator): - def can_combine( - self, function: Incomplete, argument: Incomplete - ) -> Incomplete: ... - def combine( - self, function: Incomplete, argument: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -def forwardSConstraint(left: Incomplete, right: Incomplete) -> Incomplete: ... -def backwardSxConstraint( - left: Incomplete, right: Incomplete -) -> Incomplete: ... - -ForwardSubstitution: Incomplete -BackwardSx: Incomplete - -def innermostFunction(cate: Incomplete) -> Incomplete: ... - -class UndirectedTypeRaise(UndirectedBinaryCombinator): - def can_combine( - self, function: Incomplete, arg: Incomplete - ) -> Incomplete: ... - def combine( - self, function: Incomplete, arg: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -def forwardTConstraint(left: Incomplete, right: Incomplete) -> Incomplete: ... -def backwardTConstraint(left: Incomplete, right: Incomplete) -> Incomplete: ... - -ForwardT: Incomplete -BackwardT: Incomplete diff --git a/stubs/nltk/ccg/lexicon.pyi b/stubs/nltk/ccg/lexicon.pyi deleted file mode 100644 index bfebdbd..0000000 --- a/stubs/nltk/ccg/lexicon.pyi +++ /dev/null @@ -1,61 +0,0 @@ -from _typeshed import Incomplete - -from nltk.ccg.api import CCGVar as CCGVar -from nltk.ccg.api import Direction as Direction -from nltk.ccg.api import FunctionalCategory as FunctionalCategory -from nltk.ccg.api import PrimitiveCategory as PrimitiveCategory -from nltk.internals import deprecated as deprecated -from nltk.sem.logic import Expression as Expression - -PRIM_RE: Incomplete -NEXTPRIM_RE: Incomplete -APP_RE: Incomplete -LEX_RE: Incomplete -RHS_RE: Incomplete -SEMANTICS_RE: Incomplete -COMMENTS_RE: Incomplete - -class Token: - def __init__( - self, - token: Incomplete, - categ: Incomplete, - semantics: Incomplete | None = None, - ) -> None: ... - def categ(self) -> Incomplete: ... - def semantics(self) -> Incomplete: ... - def __cmp__(self, other: Incomplete) -> Incomplete: ... - -class CCGLexicon: - def __init__( - self, - start: Incomplete, - primitives: Incomplete, - families: Incomplete, - entries: Incomplete, - ) -> None: ... - def categories(self, word: Incomplete) -> Incomplete: ... - def start(self) -> Incomplete: ... - -def matchBrackets(string: Incomplete) -> Incomplete: ... -def nextCategory(string: Incomplete) -> Incomplete: ... -def parseApplication(ap: Incomplete) -> Incomplete: ... -def parseSubscripts(subscr: Incomplete) -> Incomplete: ... -def parsePrimitiveCategory( - chunks: Incomplete, - primitives: Incomplete, - families: Incomplete, - var: Incomplete, -) -> Incomplete: ... -def augParseCategory( - line: Incomplete, - primitives: Incomplete, - families: Incomplete, - var: Incomplete | None = None, -) -> Incomplete: ... -def fromstring( - lex_str: Incomplete, include_semantics: bool = False -) -> Incomplete: ... -def parseLexicon(lex_str: Incomplete) -> Incomplete: ... - -openccg_tinytiny: Incomplete diff --git a/stubs/nltk/ccg/logic.pyi b/stubs/nltk/ccg/logic.pyi deleted file mode 100644 index e56fec8..0000000 --- a/stubs/nltk/ccg/logic.pyi +++ /dev/null @@ -1,14 +0,0 @@ -from _typeshed import Incomplete - -from nltk.sem.logic import * - -def compute_type_raised_semantics(semantics: Incomplete) -> Incomplete: ... -def compute_function_semantics( - function: Incomplete, argument: Incomplete -) -> Incomplete: ... -def compute_composition_semantics( - function: Incomplete, argument: Incomplete -) -> Incomplete: ... -def compute_substitution_semantics( - function: Incomplete, argument: Incomplete -) -> Incomplete: ... diff --git a/stubs/nltk/chat/__init__.pyi b/stubs/nltk/chat/__init__.pyi deleted file mode 100644 index a67d92f..0000000 --- a/stubs/nltk/chat/__init__.pyi +++ /dev/null @@ -1,12 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chat.eliza import eliza_chat as eliza_chat -from nltk.chat.iesha import iesha_chat as iesha_chat -from nltk.chat.rude import rude_chat as rude_chat -from nltk.chat.suntsu import suntsu_chat as suntsu_chat -from nltk.chat.util import Chat as Chat -from nltk.chat.zen import zen_chat as zen_chat - -bots: Incomplete - -def chatbots() -> None: ... diff --git a/stubs/nltk/chat/eliza.pyi b/stubs/nltk/chat/eliza.pyi deleted file mode 100644 index 8a64635..0000000 --- a/stubs/nltk/chat/eliza.pyi +++ /dev/null @@ -1,10 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chat.util import Chat as Chat -from nltk.chat.util import reflections as reflections - -pairs: Incomplete -eliza_chatbot: Incomplete - -def eliza_chat() -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/chat/iesha.pyi b/stubs/nltk/chat/iesha.pyi deleted file mode 100644 index b406cdc..0000000 --- a/stubs/nltk/chat/iesha.pyi +++ /dev/null @@ -1,10 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chat.util import Chat as Chat - -reflections: Incomplete -pairs: Incomplete -iesha_chatbot: Incomplete - -def iesha_chat() -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/chat/rude.pyi b/stubs/nltk/chat/rude.pyi deleted file mode 100644 index 88e16cd..0000000 --- a/stubs/nltk/chat/rude.pyi +++ /dev/null @@ -1,10 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chat.util import Chat as Chat -from nltk.chat.util import reflections as reflections - -pairs: Incomplete -rude_chatbot: Incomplete - -def rude_chat() -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/chat/suntsu.pyi b/stubs/nltk/chat/suntsu.pyi deleted file mode 100644 index ac18e12..0000000 --- a/stubs/nltk/chat/suntsu.pyi +++ /dev/null @@ -1,10 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chat.util import Chat as Chat -from nltk.chat.util import reflections as reflections - -pairs: Incomplete -suntsu_chatbot: Incomplete - -def suntsu_chat() -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/chat/util.pyi b/stubs/nltk/chat/util.pyi deleted file mode 100644 index e614645..0000000 --- a/stubs/nltk/chat/util.pyi +++ /dev/null @@ -1,10 +0,0 @@ -from _typeshed import Incomplete - -reflections: Incomplete - -class Chat: - def __init__( - self, pairs: Incomplete, reflections: Incomplete = {} - ) -> None: ... - def respond(self, str: Incomplete) -> Incomplete: ... - def converse(self, quit: str = "quit") -> None: ... diff --git a/stubs/nltk/chat/zen.pyi b/stubs/nltk/chat/zen.pyi deleted file mode 100644 index 0ab7cd6..0000000 --- a/stubs/nltk/chat/zen.pyi +++ /dev/null @@ -1,10 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chat.util import Chat as Chat -from nltk.chat.util import reflections as reflections - -responses: Incomplete -zen_chatbot: Incomplete - -def zen_chat() -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/chunk/__init__.pyi b/stubs/nltk/chunk/__init__.pyi deleted file mode 100644 index ad202a5..0000000 --- a/stubs/nltk/chunk/__init__.pyi +++ /dev/null @@ -1,22 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chunk.api import ChunkParserI as ChunkParserI -from nltk.chunk.named_entity import Maxent_NE_Chunker as Maxent_NE_Chunker -from nltk.chunk.regexp import RegexpChunkParser as RegexpChunkParser -from nltk.chunk.regexp import RegexpParser as RegexpParser -from nltk.chunk.util import ChunkScore as ChunkScore -from nltk.chunk.util import accuracy as accuracy -from nltk.chunk.util import conllstr2tree as conllstr2tree -from nltk.chunk.util import conlltags2tree as conlltags2tree -from nltk.chunk.util import ieerstr2tree as ieerstr2tree -from nltk.chunk.util import tagstr2tree as tagstr2tree -from nltk.chunk.util import tree2conllstr as tree2conllstr -from nltk.chunk.util import tree2conlltags as tree2conlltags - -def ne_chunker(fmt: str = "multiclass") -> Incomplete: ... -def ne_chunk( - tagged_tokens: Incomplete, binary: bool = False -) -> Incomplete: ... -def ne_chunk_sents( - tagged_sentences: Incomplete, binary: bool = False -) -> Incomplete: ... diff --git a/stubs/nltk/chunk/api.pyi b/stubs/nltk/chunk/api.pyi deleted file mode 100644 index 63ee05e..0000000 --- a/stubs/nltk/chunk/api.pyi +++ /dev/null @@ -1,10 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chunk.util import ChunkScore as ChunkScore -from nltk.internals import deprecated as deprecated -from nltk.parse import ParserI as ParserI - -class ChunkParserI(ParserI): - def parse(self, tokens: Incomplete) -> None: ... - def evaluate(self, gold: Incomplete) -> Incomplete: ... - def accuracy(self, gold: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/chunk/named_entity.pyi b/stubs/nltk/chunk/named_entity.pyi deleted file mode 100644 index b28d50c..0000000 --- a/stubs/nltk/chunk/named_entity.pyi +++ /dev/null @@ -1,41 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.chunk.api import ChunkParserI as ChunkParserI -from nltk.chunk.util import ChunkScore as ChunkScore -from nltk.classify import MaxentClassifier as MaxentClassifier -from nltk.data import find as find -from nltk.tag import ClassifierBasedTagger as ClassifierBasedTagger -from nltk.tag import pos_tag as pos_tag -from nltk.tokenize import word_tokenize as word_tokenize -from nltk.tree import Tree as Tree - -class NEChunkParserTagger(ClassifierBasedTagger): - def __init__( - self, - train: Incomplete | None = None, - classifier: Incomplete | None = None, - ) -> None: ... - -class NEChunkParser(ChunkParserI): - def __init__(self, train: Incomplete) -> None: ... - def parse(self, tokens: Incomplete) -> Incomplete: ... - -def shape(word: Incomplete) -> Incomplete: ... -def simplify_pos(s: Incomplete) -> Incomplete: ... -def postag_tree(tree: Incomplete) -> Incomplete: ... -def load_ace_data( - roots: Incomplete, fmt: str = "binary", skip_bnews: bool = True -) -> Generator[Incomplete, Incomplete, None]: ... -def load_ace_file( - textfile: Incomplete, fmt: Incomplete -) -> Generator[Incomplete, None, Incomplete]: ... -def cmp_chunks(correct: Incomplete, guessed: Incomplete) -> None: ... - -class Maxent_NE_Chunker(NEChunkParser): - def __init__(self, fmt: str = "multiclass") -> None: ... - def load_params(self) -> None: ... - def save_params(self) -> None: ... - -def build_model(fmt: str = "multiclass") -> Incomplete: ... diff --git a/stubs/nltk/chunk/regexp.pyi b/stubs/nltk/chunk/regexp.pyi deleted file mode 100644 index 3dab2df..0000000 --- a/stubs/nltk/chunk/regexp.pyi +++ /dev/null @@ -1,117 +0,0 @@ -from re import Pattern -from typing import ( - Any, - List, - Optional, - Tuple, - Union, -) - -from _typeshed import Incomplete - -from nltk.chunk.api import ChunkParserI as ChunkParserI -from nltk.tree import Tree as Tree - -class StripRule(RegexpChunkRule): - def __init__(self, tag_pattern: Incomplete, descr: Incomplete) -> None: ... - -class UnChunkRule(RegexpChunkRule): - def __init__(self, tag_pattern: Incomplete, descr: Incomplete) -> None: ... - -class MergeRule(RegexpChunkRule): - def __init__( - self, - left_tag_pattern: Incomplete, - right_tag_pattern: Incomplete, - descr: Incomplete, - ) -> None: ... - -class SplitRule(RegexpChunkRule): - def __init__( - self, - left_tag_pattern: Incomplete, - right_tag_pattern: Incomplete, - descr: Incomplete, - ) -> None: ... - -class ExpandLeftRule(RegexpChunkRule): - def __init__( - self, - left_tag_pattern: Incomplete, - right_tag_pattern: Incomplete, - descr: Incomplete, - ) -> None: ... - -class ExpandRightRule(RegexpChunkRule): - def __init__( - self, - left_tag_pattern: Incomplete, - right_tag_pattern: Incomplete, - descr: Incomplete, - ) -> None: ... - -class ChunkRuleWithContext(RegexpChunkRule): - def __init__( - self, - left_context_tag_pattern: Incomplete, - chunk_tag_pattern: Incomplete, - right_context_tag_pattern: Incomplete, - descr: Incomplete, - ) -> None: ... - -CHUNK_TAG_PATTERN: Incomplete - -def demo_eval(chunkparser: Incomplete, text: Incomplete) -> None: ... -def demo() -> None: ... -def tag_pattern2re_pattern(tag_pattern: str) -> str: ... - -class ChunkRule: - def __init__(self, tag_pattern: str, descr: str) -> None: ... - -class ChunkString: - def __init__(self, chunk_struct: Tree, debug_level: int = ...) -> None: ... - def _tag(self, tok: Tuple[str, str]) -> str: ... - def _verify(self, s: str, verify_tags: int) -> Incomplete: ... - def to_chunkstruct(self, chunk_label: str = ...) -> Tree: ... - def xform(self, regexp: Pattern, repl: str) -> Incomplete: ... # type: ignore[type-arg] - -class RegexpChunkParser: - def __init__( - self, - rules: List[ChunkRule], - chunk_label: str = ..., - root_label: str = ..., - trace: int = ..., - ) -> None: ... - def _notrace_apply(self, chunkstr: ChunkString) -> Incomplete: ... - def parse( - self, chunk_struct: List[Tuple[str, str]], trace: Optional[int] = ... - ) -> Tree: ... - -class RegexpChunkRule: - def __init__(self, regexp: Pattern, repl: str, descr: str) -> None: ... # type: ignore[type-arg] - def apply(self, chunkstr: ChunkString) -> Incomplete: ... - @staticmethod - def fromstring(s: str) -> ChunkRule: ... - -class RegexpParser: - def __init__( - self, - grammar: str, - root_label: str = ..., - loop: int = ..., - trace: int = ..., - ) -> None: ... - def _add_stage( - self, - rules: List[Union[Any, ChunkRule]], - lhs: Optional[str], - root_label: str, - trace: int, - ) -> Incomplete: ... - def _read_grammar( - self, grammar: str, root_label: str, trace: int - ) -> Incomplete: ... - def parse( - self, chunk_struct: List[Tuple[str, str]], trace: None = ... - ) -> Tree: ... diff --git a/stubs/nltk/chunk/util.pyi b/stubs/nltk/chunk/util.pyi deleted file mode 100644 index 48ec3aa..0000000 --- a/stubs/nltk/chunk/util.pyi +++ /dev/null @@ -1,58 +0,0 @@ -from _typeshed import Incomplete - -from nltk.tag.mapping import map_tag as map_tag -from nltk.tag.util import str2tuple as str2tuple -from nltk.tree import Tree as Tree - -def accuracy(chunker: Incomplete, gold: Incomplete) -> Incomplete: ... - -class ChunkScore: - def __init__(self, **kwargs: Incomplete) -> None: ... - def score(self, correct: Incomplete, guessed: Incomplete) -> None: ... - def accuracy(self) -> Incomplete: ... - def precision(self) -> Incomplete: ... - def recall(self) -> Incomplete: ... - def f_measure(self, alpha: float = 0.5) -> Incomplete: ... - def missed(self) -> Incomplete: ... - def incorrect(self) -> Incomplete: ... - def correct(self) -> Incomplete: ... - def guessed(self) -> Incomplete: ... - def __len__(self) -> int: ... - -def tagstr2tree( - s: Incomplete, - chunk_label: str = "NP", - root_label: str = "S", - sep: str = "/", - source_tagset: Incomplete | None = None, - target_tagset: Incomplete | None = None, -) -> Incomplete: ... -def conllstr2tree( - s: Incomplete, - chunk_types: Incomplete = ("NP", "PP", "VP"), - root_label: str = "S", -) -> Incomplete: ... -def tree2conlltags(t: Incomplete) -> Incomplete: ... -def conlltags2tree( - sentence: Incomplete, - chunk_types: Incomplete = ("NP", "PP", "VP"), - root_label: str = "S", - strict: bool = False, -) -> Incomplete: ... -def tree2conllstr(t: Incomplete) -> Incomplete: ... -def ieerstr2tree( - s: Incomplete, - chunk_types: Incomplete = [ - "LOCATION", - "ORGANIZATION", - "PERSON", - "DURATION", - "DATE", - "CARDINAL", - "PERCENT", - "MONEY", - "MEASURE", - ], - root_label: str = "S", -) -> Incomplete: ... -def demo() -> None: ... diff --git a/stubs/nltk/classify/__init__.pyi b/stubs/nltk/classify/__init__.pyi deleted file mode 100644 index 2666343..0000000 --- a/stubs/nltk/classify/__init__.pyi +++ /dev/null @@ -1,60 +0,0 @@ -from nltk.classify.api import ( - ClassifierI as ClassifierI, -) -from nltk.classify.api import ( - MultiClassifierI as MultiClassifierI, -) -from nltk.classify.decisiontree import ( - DecisionTreeClassifier as DecisionTreeClassifier, -) -from nltk.classify.maxent import ( - BinaryMaxentFeatureEncoding as BinaryMaxentFeatureEncoding, -) -from nltk.classify.maxent import ( - ConditionalExponentialClassifier as ConditionalExponentialClassifier, -) -from nltk.classify.maxent import ( - MaxentClassifier as MaxentClassifier, -) -from nltk.classify.maxent import ( - TypedMaxentFeatureEncoding as TypedMaxentFeatureEncoding, -) -from nltk.classify.megam import ( - call_megam as call_megam, -) -from nltk.classify.megam import ( - config_megam as config_megam, -) -from nltk.classify.naivebayes import ( - NaiveBayesClassifier as NaiveBayesClassifier, -) -from nltk.classify.positivenaivebayes import ( - PositiveNaiveBayesClassifier as PositiveNaiveBayesClassifier, -) -from nltk.classify.rte_classify import ( - RTEFeatureExtractor as RTEFeatureExtractor, -) -from nltk.classify.rte_classify import ( - rte_classifier as rte_classifier, -) -from nltk.classify.rte_classify import ( - rte_features as rte_features, -) -from nltk.classify.scikitlearn import SklearnClassifier as SklearnClassifier -from nltk.classify.senna import Senna as Senna -from nltk.classify.textcat import TextCat as TextCat -from nltk.classify.util import ( - accuracy as accuracy, -) -from nltk.classify.util import ( - apply_features as apply_features, -) -from nltk.classify.util import ( - log_likelihood as log_likelihood, -) -from nltk.classify.weka import ( - WekaClassifier as WekaClassifier, -) -from nltk.classify.weka import ( - config_weka as config_weka, -) diff --git a/stubs/nltk/classify/api.pyi b/stubs/nltk/classify/api.pyi deleted file mode 100644 index e6d4b0b..0000000 --- a/stubs/nltk/classify/api.pyi +++ /dev/null @@ -1,17 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import overridden as overridden - -class ClassifierI: - def labels(self) -> None: ... - def classify(self, featureset: Incomplete) -> Incomplete: ... - def prob_classify(self, featureset: Incomplete) -> Incomplete: ... - def classify_many(self, featuresets: Incomplete) -> Incomplete: ... - def prob_classify_many(self, featuresets: Incomplete) -> Incomplete: ... - -class MultiClassifierI: - def labels(self) -> None: ... - def classify(self, featureset: Incomplete) -> Incomplete: ... - def prob_classify(self, featureset: Incomplete) -> Incomplete: ... - def classify_many(self, featuresets: Incomplete) -> Incomplete: ... - def prob_classify_many(self, featuresets: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/classify/decisiontree.pyi b/stubs/nltk/classify/decisiontree.pyi deleted file mode 100644 index ab56cfc..0000000 --- a/stubs/nltk/classify/decisiontree.pyi +++ /dev/null @@ -1,76 +0,0 @@ -from _typeshed import Incomplete - -from nltk.classify.api import ClassifierI as ClassifierI -from nltk.probability import ( - FreqDist as FreqDist, -) -from nltk.probability import ( - MLEProbDist as MLEProbDist, -) -from nltk.probability import ( - entropy as entropy, -) - -class DecisionTreeClassifier(ClassifierI): - def __init__( - self, - label: Incomplete, - feature_name: Incomplete | None = None, - decisions: Incomplete | None = None, - default: Incomplete | None = None, - ) -> None: ... - def labels(self) -> Incomplete: ... - def classify(self, featureset: Incomplete) -> Incomplete: ... - def error(self, labeled_featuresets: Incomplete) -> Incomplete: ... - def pretty_format( - self, width: int = 70, prefix: str = "", depth: int = 4 - ) -> Incomplete: ... - def pseudocode(self, prefix: str = "", depth: int = 4) -> Incomplete: ... - @staticmethod - def train( - labeled_featuresets: Incomplete, - entropy_cutoff: float = 0.05, - depth_cutoff: int = 100, - support_cutoff: int = 10, - binary: bool = False, - feature_values: Incomplete | None = None, - verbose: bool = False, - ) -> Incomplete: ... - @staticmethod - def leaf(labeled_featuresets: Incomplete) -> Incomplete: ... - @staticmethod - def stump( - feature_name: Incomplete, labeled_featuresets: Incomplete - ) -> Incomplete: ... - def refine( - self, - labeled_featuresets: Incomplete, - entropy_cutoff: Incomplete, - depth_cutoff: Incomplete, - support_cutoff: Incomplete, - binary: bool = False, - feature_values: Incomplete | None = None, - verbose: bool = False, - ) -> None: ... - @staticmethod - def best_stump( - feature_names: Incomplete, - labeled_featuresets: Incomplete, - verbose: bool = False, - ) -> Incomplete: ... - @staticmethod - def binary_stump( - feature_name: Incomplete, - feature_value: Incomplete, - labeled_featuresets: Incomplete, - ) -> Incomplete: ... - @staticmethod - def best_binary_stump( - feature_names: Incomplete, - labeled_featuresets: Incomplete, - feature_values: Incomplete, - verbose: bool = False, - ) -> Incomplete: ... - -def f(x: Incomplete) -> Incomplete: ... -def demo() -> None: ... diff --git a/stubs/nltk/classify/maxent.pyi b/stubs/nltk/classify/maxent.pyi deleted file mode 100644 index a87149f..0000000 --- a/stubs/nltk/classify/maxent.pyi +++ /dev/null @@ -1,84 +0,0 @@ -from typing import ( - Dict, - List, - Optional, - Set, - Tuple, - Union, -) - -from _typeshed import Incomplete - -from nltk.collections import OrderedDict # type: ignore[import-untyped] - -def train_maxent_classifier_with_megam( - train_toks: List[Tuple[Dict[str, int], str]], - trace: int = ..., - encoding: None = ..., - labels: None = ..., - gaussian_prior_sigma: int = ..., - **kwargs: Incomplete, -) -> Incomplete: ... - -class BinaryMaxentFeatureEncoding: - def __init__( - self, - labels: Union[List[str], Set[str]], - mapping: Union[Dict[Tuple[str, int, str], int], OrderedDict], - unseen_features: bool = ..., - alwayson_features: bool = ..., - ) -> None: ... - def encode( - self, featureset: Dict[str, int], label: str - ) -> List[Tuple[int, int]]: ... - def labels(self) -> List[str]: ... - @classmethod - def train( - cls: Incomplete, - train_toks: List[Tuple[Dict[str, int], str]], - count_cutoff: int = ..., - labels: None = ..., - **options: Incomplete, - ) -> BinaryMaxentFeatureEncoding: ... - -class MaxentClassifier: - @classmethod - def train( - cls: Incomplete, - train_toks: List[Tuple[Dict[str, int], str]], - algorithm: Optional[str] = ..., - trace: int = ..., - encoding: None = ..., - labels: None = ..., - gaussian_prior_sigma: int = ..., - **cutoffs: Incomplete, - ) -> Incomplete: ... - -class TadmEventMaxentFeatureEncoding: - def __init__( - self, - labels: List[str], - mapping: OrderedDict, - unseen_features: bool = ..., - alwayson_features: bool = ..., - ) -> None: ... - def encode( - self, featureset: Dict[str, int], label: str - ) -> List[Tuple[int, int]]: ... - def labels(self) -> List[str]: ... - @classmethod - def train( - cls: Incomplete, - train_toks: List[Tuple[Dict[str, int], str]], - count_cutoff: int = ..., - labels: None = ..., - **options: Incomplete, - ) -> TadmEventMaxentFeatureEncoding: ... - -class TadmMaxentClassifier: - @classmethod - def train( - cls: Incomplete, - train_toks: List[Tuple[Dict[str, int], str]], - **kwargs: Incomplete, - ) -> Incomplete: ... diff --git a/stubs/nltk/classify/megam.pyi b/stubs/nltk/classify/megam.pyi deleted file mode 100644 index b9fc442..0000000 --- a/stubs/nltk/classify/megam.pyi +++ /dev/null @@ -1,23 +0,0 @@ -from io import TextIOWrapper -from typing import ( - Dict, - List, - Tuple, -) - -from _typeshed import Incomplete - -from nltk.classify.maxent import BinaryMaxentFeatureEncoding - -def _write_megam_features( - vector: List[Tuple[int, int]], stream: TextIOWrapper, bernoulli: bool -) -> Incomplete: ... -def call_megam(args: List[str]) -> Incomplete: ... -def config_megam(bin: None = ...) -> Incomplete: ... -def write_megam_file( - train_toks: List[Tuple[Dict[str, int], str]], - encoding: BinaryMaxentFeatureEncoding, - stream: TextIOWrapper, - bernoulli: bool = ..., - explicit: bool = ..., -) -> Incomplete: ... diff --git a/stubs/nltk/classify/naivebayes.pyi b/stubs/nltk/classify/naivebayes.pyi deleted file mode 100644 index a21ce7a..0000000 --- a/stubs/nltk/classify/naivebayes.pyi +++ /dev/null @@ -1,33 +0,0 @@ -from _typeshed import Incomplete - -from nltk.classify.api import ClassifierI as ClassifierI -from nltk.probability import ( - DictionaryProbDist as DictionaryProbDist, -) -from nltk.probability import ( - ELEProbDist as ELEProbDist, -) -from nltk.probability import ( - FreqDist as FreqDist, -) -from nltk.probability import ( - sum_logs as sum_logs, -) - -class NaiveBayesClassifier(ClassifierI): - def __init__( - self, label_probdist: Incomplete, feature_probdist: Incomplete - ) -> None: ... - def labels(self) -> Incomplete: ... - def classify(self, featureset: Incomplete) -> Incomplete: ... - def prob_classify(self, featureset: Incomplete) -> Incomplete: ... - def show_most_informative_features(self, n: int = 10) -> Incomplete: ... - def most_informative_features(self, n: int = 100) -> Incomplete: ... - @classmethod - def train( - cls: Incomplete, - labeled_featuresets: Incomplete, - estimator: Incomplete = ..., - ) -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/classify/positivenaivebayes.pyi b/stubs/nltk/classify/positivenaivebayes.pyi deleted file mode 100644 index 208c472..0000000 --- a/stubs/nltk/classify/positivenaivebayes.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -from nltk.classify.naivebayes import ( - NaiveBayesClassifier as NaiveBayesClassifier, -) -from nltk.probability import ( - DictionaryProbDist as DictionaryProbDist, -) -from nltk.probability import ( - ELEProbDist as ELEProbDist, -) -from nltk.probability import ( - FreqDist as FreqDist, -) - -class PositiveNaiveBayesClassifier(NaiveBayesClassifier): - @staticmethod - def train( - positive_featuresets: Incomplete, - unlabeled_featuresets: Incomplete, - positive_prob_prior: float = 0.5, - estimator: Incomplete = ..., - ) -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/classify/rte_classify.pyi b/stubs/nltk/classify/rte_classify.pyi deleted file mode 100644 index cea9444..0000000 --- a/stubs/nltk/classify/rte_classify.pyi +++ /dev/null @@ -1,7 +0,0 @@ -from typing import Optional - -from _typeshed import Incomplete - -def rte_classifier( - algorithm: str, sample_N: Optional[int] = ... -) -> Incomplete: ... diff --git a/stubs/nltk/classify/scikitlearn.pyi b/stubs/nltk/classify/scikitlearn.pyi deleted file mode 100644 index 2537915..0000000 --- a/stubs/nltk/classify/scikitlearn.pyi +++ /dev/null @@ -1,17 +0,0 @@ -from _typeshed import Incomplete - -from nltk.classify.api import ClassifierI - -__all__ = ["SklearnClassifier"] - -class SklearnClassifier(ClassifierI): - def __init__( - self, - estimator: Incomplete, - dtype: Incomplete = ..., - sparse: bool = True, - ) -> None: ... - def classify_many(self, featuresets: Incomplete) -> Incomplete: ... - def prob_classify_many(self, featuresets: Incomplete) -> Incomplete: ... - def labels(self) -> Incomplete: ... - def train(self, labeled_featuresets: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/classify/senna.pyi b/stubs/nltk/classify/senna.pyi deleted file mode 100644 index af6b541..0000000 --- a/stubs/nltk/classify/senna.pyi +++ /dev/null @@ -1,11 +0,0 @@ -from typing import List - -from _typeshed import Incomplete - -class Senna: - def __init__( - self, senna_path: str, operations: List[str], encoding: str = ... - ) -> None: ... - def executable(self, base_path: str) -> str: ... - def tag(self, tokens: List[str]) -> Incomplete: ... - def tag_sents(self, sentences: List[List[str]]) -> Incomplete: ... diff --git a/stubs/nltk/classify/tadm.pyi b/stubs/nltk/classify/tadm.pyi deleted file mode 100644 index cbbe01e..0000000 --- a/stubs/nltk/classify/tadm.pyi +++ /dev/null @@ -1,18 +0,0 @@ -from io import TextIOWrapper -from typing import ( - Dict, - List, - Tuple, -) - -from _typeshed import Incomplete - -from nltk.classify.maxent import TadmEventMaxentFeatureEncoding - -def call_tadm(args: List[str]) -> Incomplete: ... -def config_tadm(bin: None = ...) -> Incomplete: ... -def write_tadm_file( - train_toks: List[Tuple[Dict[str, int], str]], - encoding: TadmEventMaxentFeatureEncoding, - stream: TextIOWrapper, -) -> Incomplete: ... diff --git a/stubs/nltk/classify/textcat.pyi b/stubs/nltk/classify/textcat.pyi deleted file mode 100644 index 0d9a2a0..0000000 --- a/stubs/nltk/classify/textcat.pyi +++ /dev/null @@ -1,17 +0,0 @@ -from _typeshed import Incomplete - -from nltk.util import trigrams as trigrams - -class TextCat: - fingerprints: Incomplete - last_distances: Incomplete - def __init__(self) -> None: ... - def remove_punctuation(self, text: Incomplete) -> Incomplete: ... - def profile(self, text: Incomplete) -> Incomplete: ... - def calc_dist( - self, lang: Incomplete, trigram: Incomplete, text_profile: Incomplete - ) -> Incomplete: ... - def lang_dists(self, text: Incomplete) -> Incomplete: ... - def guess_language(self, text: Incomplete) -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/classify/util.pyi b/stubs/nltk/classify/util.pyi deleted file mode 100644 index 578c9ea..0000000 --- a/stubs/nltk/classify/util.pyi +++ /dev/null @@ -1,35 +0,0 @@ -from _typeshed import Incomplete - -from nltk.util import LazyMap as LazyMap - -def apply_features( - feature_func: Incomplete, - toks: Incomplete, - labeled: Incomplete | None = None, -) -> Incomplete: ... -def attested_labels(tokens: Incomplete) -> Incomplete: ... -def log_likelihood(classifier: Incomplete, gold: Incomplete) -> Incomplete: ... -def accuracy(classifier: Incomplete, gold: Incomplete) -> Incomplete: ... - -class CutoffChecker: - cutoffs: Incomplete - ll: Incomplete - acc: Incomplete - iter: int - def __init__(self, cutoffs: Incomplete) -> None: ... - def check( - self, classifier: Incomplete, train_toks: Incomplete - ) -> Incomplete: ... - -def names_demo_features(name: Incomplete) -> Incomplete: ... -def binary_names_demo_features(name: Incomplete) -> Incomplete: ... -def names_demo( - trainer: Incomplete, features: Incomplete = ... -) -> Incomplete: ... -def partial_names_demo( - trainer: Incomplete, features: Incomplete = ... -) -> Incomplete: ... -def wsd_demo( - trainer: Incomplete, word: Incomplete, features: Incomplete, n: int = 1000 -) -> Incomplete: ... -def check_megam_config() -> None: ... diff --git a/stubs/nltk/classify/weka.pyi b/stubs/nltk/classify/weka.pyi deleted file mode 100644 index 1c81c7d..0000000 --- a/stubs/nltk/classify/weka.pyi +++ /dev/null @@ -1,38 +0,0 @@ -from _typeshed import Incomplete - -from nltk.classify.api import ClassifierI as ClassifierI -from nltk.internals import config_java as config_java -from nltk.internals import java as java -from nltk.probability import DictionaryProbDist as DictionaryProbDist - -def config_weka(classpath: Incomplete | None = None) -> None: ... - -class WekaClassifier(ClassifierI): - def __init__( - self, formatter: Incomplete, model_filename: Incomplete - ) -> None: ... - def prob_classify_many(self, featuresets: Incomplete) -> Incomplete: ... - def classify_many(self, featuresets: Incomplete) -> Incomplete: ... - def parse_weka_distribution(self, s: Incomplete) -> Incomplete: ... - def parse_weka_output(self, lines: Incomplete) -> Incomplete: ... - @classmethod - def train( - cls: Incomplete, - model_filename: Incomplete, - featuresets: Incomplete, - classifier: str = "naivebayes", - options: Incomplete = [], - quiet: bool = True, - ) -> Incomplete: ... - -class ARFF_Formatter: - def __init__(self, labels: Incomplete, features: Incomplete) -> None: ... - def format(self, tokens: Incomplete) -> Incomplete: ... - def labels(self) -> Incomplete: ... - def write(self, outfile: Incomplete, tokens: Incomplete) -> None: ... - @staticmethod - def from_train(tokens: Incomplete) -> Incomplete: ... - def header_section(self) -> Incomplete: ... - def data_section( - self, tokens: Incomplete, labeled: Incomplete | None = None - ) -> Incomplete: ... diff --git a/stubs/nltk/collocations.pyi b/stubs/nltk/collocations.pyi deleted file mode 100644 index a272750..0000000 --- a/stubs/nltk/collocations.pyi +++ /dev/null @@ -1,97 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -__all__ = [ - "BigramCollocationFinder", - "TrigramCollocationFinder", - "QuadgramCollocationFinder", -] - -class AbstractCollocationFinder: - word_fd: Incomplete - N: Incomplete - ngram_fd: Incomplete - def __init__(self, word_fd: Incomplete, ngram_fd: Incomplete) -> None: ... - @classmethod - def from_documents( - cls: Incomplete, documents: Incomplete - ) -> Incomplete: ... - def apply_freq_filter(self, min_freq: Incomplete) -> Incomplete: ... - def apply_ngram_filter(self, fn: Incomplete) -> Incomplete: ... - def apply_word_filter(self, fn: Incomplete) -> Incomplete: ... - def score_ngrams(self, score_fn: Incomplete) -> Incomplete: ... - def nbest(self, score_fn: Incomplete, n: Incomplete) -> Incomplete: ... - def above_score( - self, score_fn: Incomplete, min_score: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class BigramCollocationFinder(AbstractCollocationFinder): - default_ws: int - window_size: Incomplete - def __init__( - self, word_fd: Incomplete, bigram_fd: Incomplete, window_size: int = 2 - ) -> None: ... - @classmethod - def from_words( - cls: Incomplete, words: Incomplete, window_size: int = 2 - ) -> Incomplete: ... - def score_ngram( - self, score_fn: Incomplete, w1: Incomplete, w2: Incomplete - ) -> Incomplete: ... - -class TrigramCollocationFinder(AbstractCollocationFinder): - default_ws: int - wildcard_fd: Incomplete - bigram_fd: Incomplete - def __init__( - self, - word_fd: Incomplete, - bigram_fd: Incomplete, - wildcard_fd: Incomplete, - trigram_fd: Incomplete, - ) -> None: ... - @classmethod - def from_words( - cls: Incomplete, words: Incomplete, window_size: int = 3 - ) -> Incomplete: ... - def bigram_finder(self) -> Incomplete: ... - def score_ngram( - self, - score_fn: Incomplete, - w1: Incomplete, - w2: Incomplete, - w3: Incomplete, - ) -> Incomplete: ... - -class QuadgramCollocationFinder(AbstractCollocationFinder): - default_ws: int - iii: Incomplete - ii: Incomplete - ixi: Incomplete - ixxi: Incomplete - iixi: Incomplete - ixii: Incomplete - def __init__( - self, - word_fd: Incomplete, - quadgram_fd: Incomplete, - ii: Incomplete, - iii: Incomplete, - ixi: Incomplete, - ixxi: Incomplete, - iixi: Incomplete, - ixii: Incomplete, - ) -> None: ... - @classmethod - def from_words( - cls: Incomplete, words: Incomplete, window_size: int = 4 - ) -> Incomplete: ... - def score_ngram( - self, - score_fn: Incomplete, - w1: Incomplete, - w2: Incomplete, - w3: Incomplete, - w4: Incomplete, - ) -> Incomplete: ... diff --git a/stubs/nltk/compat.pyi b/stubs/nltk/compat.pyi deleted file mode 100644 index 28e0acb..0000000 --- a/stubs/nltk/compat.pyi +++ /dev/null @@ -1,6 +0,0 @@ -from _typeshed import Incomplete - -DATA_UPDATES: Incomplete - -def add_py3_data(path: Incomplete) -> Incomplete: ... -def py3_data(init_func: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/__init__.pyi b/stubs/nltk/corpus/__init__.pyi index 44c33b3..1dd3f63 100644 --- a/stubs/nltk/corpus/__init__.pyi +++ b/stubs/nltk/corpus/__init__.pyi @@ -1,86 +1,3 @@ from nltk.corpus.reader import * -from nltk.corpus.util import LazyCorpusLoader as LazyCorpusLoader -from nltk.tokenize import RegexpTokenizer as RegexpTokenizer -abc: PlaintextCorpusReader -alpino: AlpinoCorpusReader -bcp47: BCP47CorpusReader -brown: CategorizedTaggedCorpusReader -cess_cat: BracketParseCorpusReader -cess_esp: BracketParseCorpusReader -cmudict: CMUDictCorpusReader -comtrans: AlignedCorpusReader -comparative_sentences: ComparativeSentencesCorpusReader -conll2000: ConllChunkCorpusReader -conll2002: ConllChunkCorpusReader -conll2007: DependencyCorpusReader -crubadan: CrubadanCorpusReader -dependency_treebank: DependencyCorpusReader -extended_omw: CorpusReader -floresta: BracketParseCorpusReader -framenet15: FramenetCorpusReader -framenet: FramenetCorpusReader -gazetteers: WordListCorpusReader -genesis: PlaintextCorpusReader -gutenberg: PlaintextCorpusReader -ieer: IEERCorpusReader -inaugural: PlaintextCorpusReader -indian: IndianCorpusReader -jeita: ChasenCorpusReader -knbc: KNBCorpusReader -lin_thesaurus: LinThesaurusCorpusReader -mac_morpho: MacMorphoCorpusReader -machado: PortugueseCategorizedPlaintextCorpusReader -masc_tagged: CategorizedTaggedCorpusReader -movie_reviews: CategorizedPlaintextCorpusReader -multext_east: MTECorpusReader -names: WordListCorpusReader -nps_chat: NPSChatCorpusReader -opinion_lexicon: OpinionLexiconCorpusReader -ppattach: PPAttachmentCorpusReader -product_reviews_1: ReviewsCorpusReader -product_reviews_2: ReviewsCorpusReader -pros_cons: ProsConsCorpusReader -ptb: CategorizedBracketParseCorpusReader -qc: StringCategoryCorpusReader -reuters: CategorizedPlaintextCorpusReader -rte: RTECorpusReader -senseval: SensevalCorpusReader -sentence_polarity: CategorizedSentencesCorpusReader -sentiwordnet: SentiWordNetCorpusReader -shakespeare: XMLCorpusReader -sinica_treebank: SinicaTreebankCorpusReader -state_union: PlaintextCorpusReader -stopwords: WordListCorpusReader -subjectivity: CategorizedSentencesCorpusReader -swadesh: SwadeshCorpusReader -swadesh110: PanlexSwadeshCorpusReader -swadesh207: PanlexSwadeshCorpusReader -switchboard: SwitchboardCorpusReader -timit: TimitCorpusReader -timit_tagged: TimitTaggedCorpusReader -toolbox: ToolboxCorpusReader -treebank: BracketParseCorpusReader -treebank_chunk: ChunkedCorpusReader -treebank_raw: PlaintextCorpusReader -twitter_samples: TwitterCorpusReader -udhr: UdhrCorpusReader -udhr2: PlaintextCorpusReader -universal_treebanks: ConllCorpusReader -verbnet: VerbnetCorpusReader -webtext: PlaintextCorpusReader wordnet: WordNetCorpusReader -wordnet31: WordNetCorpusReader -wordnet2021: WordNetCorpusReader -wordnet2022: WordNetCorpusReader -wordnet_ic: WordNetICCorpusReader -words: WordListCorpusReader -propbank: PropbankCorpusReader -nombank: NombankCorpusReader -propbank_ptb: PropbankCorpusReader -nombank_ptb: NombankCorpusReader -semcor: SemcorCorpusReader -nonbreaking_prefixes: NonbreakingPrefixesCorpusReader -perluniprops: UnicharsCorpusReader - -def demo() -> None: ... diff --git a/stubs/nltk/corpus/europarl_raw.pyi b/stubs/nltk/corpus/europarl_raw.pyi deleted file mode 100644 index 019b405..0000000 --- a/stubs/nltk/corpus/europarl_raw.pyi +++ /dev/null @@ -1,14 +0,0 @@ -from nltk.corpus.reader import * -from nltk.corpus.util import LazyCorpusLoader as LazyCorpusLoader - -danish: EuroparlCorpusReader -dutch: EuroparlCorpusReader -english: EuroparlCorpusReader -finnish: EuroparlCorpusReader -french: EuroparlCorpusReader -german: EuroparlCorpusReader -greek: EuroparlCorpusReader -italian: EuroparlCorpusReader -portuguese: EuroparlCorpusReader -spanish: EuroparlCorpusReader -swedish: EuroparlCorpusReader diff --git a/stubs/nltk/corpus/reader/__init__.pyi b/stubs/nltk/corpus/reader/__init__.pyi index 6e8f36b..7311625 100644 --- a/stubs/nltk/corpus/reader/__init__.pyi +++ b/stubs/nltk/corpus/reader/__init__.pyi @@ -1,198 +1 @@ -from nltk.corpus.reader.aligned import * -from nltk.corpus.reader.api import * -from nltk.corpus.reader.bcp47 import * -from nltk.corpus.reader.bnc import * -from nltk.corpus.reader.bracket_parse import * -from nltk.corpus.reader.categorized_sents import * -from nltk.corpus.reader.chasen import * -from nltk.corpus.reader.childes import * -from nltk.corpus.reader.chunked import * -from nltk.corpus.reader.cmudict import * -from nltk.corpus.reader.comparative_sents import * -from nltk.corpus.reader.conll import * -from nltk.corpus.reader.crubadan import * -from nltk.corpus.reader.dependency import * -from nltk.corpus.reader.framenet import * -from nltk.corpus.reader.ieer import * -from nltk.corpus.reader.indian import * -from nltk.corpus.reader.ipipan import * -from nltk.corpus.reader.knbc import * -from nltk.corpus.reader.lin import * -from nltk.corpus.reader.mte import * -from nltk.corpus.reader.nkjp import * -from nltk.corpus.reader.nombank import * -from nltk.corpus.reader.nps_chat import * -from nltk.corpus.reader.opinion_lexicon import * -from nltk.corpus.reader.panlex_lite import * -from nltk.corpus.reader.panlex_swadesh import * -from nltk.corpus.reader.pl196x import * -from nltk.corpus.reader.plaintext import * -from nltk.corpus.reader.ppattach import * -from nltk.corpus.reader.propbank import * -from nltk.corpus.reader.pros_cons import * -from nltk.corpus.reader.reviews import * -from nltk.corpus.reader.rte import * -from nltk.corpus.reader.semcor import * -from nltk.corpus.reader.senseval import * -from nltk.corpus.reader.sentiwordnet import * -from nltk.corpus.reader.sinica_treebank import * -from nltk.corpus.reader.string_category import * -from nltk.corpus.reader.switchboard import * -from nltk.corpus.reader.tagged import * -from nltk.corpus.reader.timit import * -from nltk.corpus.reader.toolbox import * -from nltk.corpus.reader.twitter import * -from nltk.corpus.reader.udhr import * -from nltk.corpus.reader.util import * -from nltk.corpus.reader.verbnet import * -from nltk.corpus.reader.wordlist import * from nltk.corpus.reader.wordnet import * -from nltk.corpus.reader.xmldocs import * -from nltk.corpus.reader.ycoe import * - -__all__ = [ - "CorpusReader", - "CategorizedCorpusReader", - "PlaintextCorpusReader", - "find_corpus_fileids", - "TaggedCorpusReader", - "CMUDictCorpusReader", - "ConllChunkCorpusReader", - "WordListCorpusReader", - "PPAttachmentCorpusReader", - "SensevalCorpusReader", - "IEERCorpusReader", - "ChunkedCorpusReader", - "SinicaTreebankCorpusReader", - "BracketParseCorpusReader", - "IndianCorpusReader", - "ToolboxCorpusReader", - "TimitCorpusReader", - "YCOECorpusReader", - "MacMorphoCorpusReader", - "SyntaxCorpusReader", - "AlpinoCorpusReader", - "RTECorpusReader", - "StringCategoryCorpusReader", - "EuroparlCorpusReader", - "CategorizedBracketParseCorpusReader", - "CategorizedTaggedCorpusReader", - "CategorizedPlaintextCorpusReader", - "PortugueseCategorizedPlaintextCorpusReader", - "tagged_treebank_para_block_reader", - "PropbankCorpusReader", - "VerbnetCorpusReader", - "BNCCorpusReader", - "ConllCorpusReader", - "XMLCorpusReader", - "NPSChatCorpusReader", - "SwadeshCorpusReader", - "WordNetCorpusReader", - "WordNetICCorpusReader", - "SwitchboardCorpusReader", - "DependencyCorpusReader", - "NombankCorpusReader", - "IPIPANCorpusReader", - "Pl196xCorpusReader", - "TEICorpusView", - "KNBCorpusReader", - "ChasenCorpusReader", - "CHILDESCorpusReader", - "AlignedCorpusReader", - "TimitTaggedCorpusReader", - "LinThesaurusCorpusReader", - "SemcorCorpusReader", - "FramenetCorpusReader", - "UdhrCorpusReader", - "BNCCorpusReader", - "SentiWordNetCorpusReader", - "SentiSynset", - "TwitterCorpusReader", - "NKJPCorpusReader", - "CrubadanCorpusReader", - "MTECorpusReader", - "ReviewsCorpusReader", - "OpinionLexiconCorpusReader", - "ProsConsCorpusReader", - "CategorizedSentencesCorpusReader", - "ComparativeSentencesCorpusReader", - "PanLexLiteCorpusReader", - "NonbreakingPrefixesCorpusReader", - "UnicharsCorpusReader", - "MWAPPDBCorpusReader", - "PanlexSwadeshCorpusReader", - "BCP47CorpusReader", -] - -# Names in __all__ with no definition: -# AlignedCorpusReader -# AlpinoCorpusReader -# BCP47CorpusReader -# BNCCorpusReader -# BNCCorpusReader -# BracketParseCorpusReader -# CHILDESCorpusReader -# CMUDictCorpusReader -# CategorizedBracketParseCorpusReader -# CategorizedCorpusReader -# CategorizedPlaintextCorpusReader -# CategorizedSentencesCorpusReader -# CategorizedTaggedCorpusReader -# ChasenCorpusReader -# ChunkedCorpusReader -# ComparativeSentencesCorpusReader -# ConllChunkCorpusReader -# ConllCorpusReader -# CorpusReader -# CrubadanCorpusReader -# DependencyCorpusReader -# EuroparlCorpusReader -# FramenetCorpusReader -# IEERCorpusReader -# IPIPANCorpusReader -# IndianCorpusReader -# KNBCorpusReader -# LinThesaurusCorpusReader -# MTECorpusReader -# MWAPPDBCorpusReader -# MacMorphoCorpusReader -# NKJPCorpusReader -# NPSChatCorpusReader -# NombankCorpusReader -# NonbreakingPrefixesCorpusReader -# OpinionLexiconCorpusReader -# PPAttachmentCorpusReader -# PanLexLiteCorpusReader -# PanlexSwadeshCorpusReader -# Pl196xCorpusReader -# PlaintextCorpusReader -# PortugueseCategorizedPlaintextCorpusReader -# PropbankCorpusReader -# ProsConsCorpusReader -# RTECorpusReader -# ReviewsCorpusReader -# SemcorCorpusReader -# SensevalCorpusReader -# SentiSynset -# SentiWordNetCorpusReader -# SinicaTreebankCorpusReader -# StringCategoryCorpusReader -# SwadeshCorpusReader -# SwitchboardCorpusReader -# SyntaxCorpusReader -# TEICorpusView -# TaggedCorpusReader -# TimitCorpusReader -# TimitTaggedCorpusReader -# ToolboxCorpusReader -# TwitterCorpusReader -# UdhrCorpusReader -# UnicharsCorpusReader -# VerbnetCorpusReader -# WordListCorpusReader -# WordNetCorpusReader -# WordNetICCorpusReader -# XMLCorpusReader -# YCOECorpusReader -# find_corpus_fileids -# tagged_treebank_para_block_reader diff --git a/stubs/nltk/corpus/reader/aligned.pyi b/stubs/nltk/corpus/reader/aligned.pyi deleted file mode 100644 index 1251a9b..0000000 --- a/stubs/nltk/corpus/reader/aligned.pyi +++ /dev/null @@ -1,50 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import CorpusReader as CorpusReader -from nltk.corpus.reader.util import ( - StreamBackedCorpusView as StreamBackedCorpusView, -) -from nltk.corpus.reader.util import ( - concat as concat, -) -from nltk.corpus.reader.util import ( - read_alignedsent_block as read_alignedsent_block, -) -from nltk.tokenize import ( - RegexpTokenizer as RegexpTokenizer, -) -from nltk.tokenize import ( - WhitespaceTokenizer as WhitespaceTokenizer, -) -from nltk.translate import AlignedSent as AlignedSent -from nltk.translate import Alignment as Alignment - -class AlignedCorpusReader(CorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - sep: str = "/", - word_tokenizer: Incomplete = ..., - sent_tokenizer: Incomplete = ..., - alignedsent_block_reader: Incomplete = ..., - encoding: str = "latin1", - ) -> None: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def aligned_sents( - self, fileids: Incomplete | None = None - ) -> Incomplete: ... - -class AlignedSentCorpusView(StreamBackedCorpusView): - def __init__( - self, - corpus_file: Incomplete, - encoding: Incomplete, - aligned: Incomplete, - group_by_sent: Incomplete, - word_tokenizer: Incomplete, - sent_tokenizer: Incomplete, - alignedsent_block_reader: Incomplete, - ) -> None: ... - def read_block(self, stream: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/api.pyi b/stubs/nltk/corpus/reader/api.pyi index 233ee1f..17c1b28 100644 --- a/stubs/nltk/corpus/reader/api.pyi +++ b/stubs/nltk/corpus/reader/api.pyi @@ -1,87 +1,33 @@ -from typing import ( - Any, - Dict, - List, - Optional, - Tuple, - Union, -) +from typing import Any from _typeshed import Incomplete -from nltk.corpus.reader.util import ( - ConcatenatedCorpusView, -) -from nltk.corpus.reader.util import ( - StreamBackedCorpusView as StreamBackedCorpusView, -) from nltk.data import ( FileSystemPathPointer, SeekableUnicodeStreamReader, ZipFilePathPointer, ) -class CategorizedCorpusReader: - def __init__(self, kwargs: Dict[str, str]) -> Incomplete: ... - def _init(self) -> Incomplete: ... - def _resolve( - self, fileids: Optional[str], categories: Optional[List[str]] - ) -> str: ... - def categories(self, fileids: None = ...) -> Incomplete: ... - def fileids( - self, categories: Optional[Union[List[str], str]] = ... - ) -> List[Any]: ... - def words( - self, - fileids: Optional[str] = ..., - categories: Optional[List[str]] = ..., - ) -> Incomplete: ... - class CorpusReader: def __init__( self, - root: Union[ZipFilePathPointer, FileSystemPathPointer], - fileids: Union[ - str, - Tuple[str, ...], - ], - encoding: Union[str, List[Tuple[str, str]]] = ..., + root: ZipFilePathPointer | FileSystemPathPointer, + fileids: str | tuple[str, ...], + encoding: str | list[tuple[str, str]] = ..., tagset: None = ..., ) -> Incomplete: ... def abspath(self, fileid: str) -> ZipFilePathPointer: ... def abspaths( self, - fileids: Optional[str] = ..., + fileids: str | None = ..., include_encoding: bool = ..., include_fileid: bool = ..., - ) -> List[ - Union[ - Tuple[FileSystemPathPointer, str], - Tuple[ZipFilePathPointer, str], - Any, - ] + ) -> list[ + tuple[FileSystemPathPointer, str] + | tuple[ZipFilePathPointer, str] + | Any ]: ... def encoding(self, file: str) -> str: ... def ensure_loaded(self) -> Incomplete: ... - def fileids(self) -> List[Any]: ... + def fileids(self) -> list[Any]: ... def open(self, file: str) -> SeekableUnicodeStreamReader: ... - -class SyntaxCorpusReader: - def _read_sent_block( - self, stream: SeekableUnicodeStreamReader - ) -> List[List[str]]: ... - def _read_tagged_sent_block( - self, stream: SeekableUnicodeStreamReader, tagset: None = ... - ) -> List[List[Tuple[str, str]]]: ... - def _read_word_block( - self, stream: SeekableUnicodeStreamReader - ) -> List[str]: ... - def tagged_sents( - self, fileids: None = ..., tagset: None = ... - ) -> ConcatenatedCorpusView: ... - def tagged_words( - self, fileids: Optional[str] = ..., tagset: None = ... - ) -> Incomplete: ... - def words( - self, fileids: Optional[str] = ... - ) -> Union[ConcatenatedCorpusView, StreamBackedCorpusView]: ... diff --git a/stubs/nltk/corpus/reader/bcp47.pyi b/stubs/nltk/corpus/reader/bcp47.pyi deleted file mode 100644 index 7893ddc..0000000 --- a/stubs/nltk/corpus/reader/bcp47.pyi +++ /dev/null @@ -1,22 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader import CorpusReader as CorpusReader - -class BCP47CorpusReader(CorpusReader): - langcode: Incomplete - db: Incomplete - subdiv: Incomplete - def __init__(self, root: Incomplete, fileids: Incomplete) -> None: ... - wiki_q: Incomplete - def load_wiki_q(self) -> None: ... - def wiki_dict(self, lines: Incomplete) -> Incomplete: ... - def subdiv_dict(self, subdivs: Incomplete) -> Incomplete: ... - casing: Incomplete - format: Incomplete - def morphology(self) -> None: ... - version: Incomplete - def data_dict(self, records: Incomplete) -> Incomplete: ... - def val2str(self, val: Incomplete) -> Incomplete: ... - def lang2str(self, lg_record: Incomplete) -> Incomplete: ... - def parse_tag(self, tag: Incomplete) -> Incomplete: ... - def name(self, tag: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/bnc.pyi b/stubs/nltk/corpus/reader/bnc.pyi deleted file mode 100644 index af10347..0000000 --- a/stubs/nltk/corpus/reader/bnc.pyi +++ /dev/null @@ -1,68 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.util import concat as concat -from nltk.corpus.reader.xmldocs import ( - ElementTree as ElementTree, -) -from nltk.corpus.reader.xmldocs import ( - XMLCorpusReader as XMLCorpusReader, -) -from nltk.corpus.reader.xmldocs import ( - XMLCorpusView as XMLCorpusView, -) - -class BNCCorpusReader(XMLCorpusReader): - def __init__( - self, root: Incomplete, fileids: Incomplete, lazy: bool = True - ) -> None: ... - def words( - self, - fileids: Incomplete | None = None, - strip_space: bool = True, - stem: bool = False, - ) -> Incomplete: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - c5: bool = False, - strip_space: bool = True, - stem: bool = False, - ) -> Incomplete: ... - def sents( - self, - fileids: Incomplete | None = None, - strip_space: bool = True, - stem: bool = False, - ) -> Incomplete: ... - def tagged_sents( - self, - fileids: Incomplete | None = None, - c5: bool = False, - strip_space: bool = True, - stem: bool = False, - ) -> Incomplete: ... - -class BNCSentence(list): - num: Incomplete - def __init__(self, num: Incomplete, items: Incomplete) -> None: ... - -class BNCWordView(XMLCorpusView): - tags_to_ignore: Incomplete - title: Incomplete - author: Incomplete - editor: Incomplete - resps: Incomplete - def __init__( - self, - fileid: Incomplete, - sent: Incomplete, - tag: Incomplete, - strip_space: Incomplete, - stem: Incomplete, - ) -> None: ... - def handle_header(self, elt: Incomplete, context: Incomplete) -> None: ... - def handle_elt( - self, elt: Incomplete, context: Incomplete - ) -> Incomplete: ... - def handle_word(self, elt: Incomplete) -> Incomplete: ... - def handle_sent(self, elt: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/bracket_parse.pyi b/stubs/nltk/corpus/reader/bracket_parse.pyi deleted file mode 100644 index 9bc5b45..0000000 --- a/stubs/nltk/corpus/reader/bracket_parse.pyi +++ /dev/null @@ -1,48 +0,0 @@ -from typing import ( - List, - Optional, - Tuple, -) - -from _typeshed import Incomplete - -from nltk.data import ( - FileSystemPathPointer, - SeekableUnicodeStreamReader, -) - -class BracketParseCorpusReader: - def __init__( - self, - root: FileSystemPathPointer, - fileids: str, - comment_char: Optional[str] = ..., - detect_blocks: str = ..., - encoding: str = ..., - tagset: Optional[str] = ..., - ) -> Incomplete: ... - def _normalize(self, t: str) -> str: ... - def _read_block( - self, stream: SeekableUnicodeStreamReader - ) -> List[str]: ... - def _tag(self, t: str, tagset: None = ...) -> List[Tuple[str, str]]: ... - def _word(self, t: str) -> List[str]: ... - -class CategorizedBracketParseCorpusReader: - def __init__( - self, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def tagged_words( - self, - fileids: Optional[str] = ..., - categories: None = ..., - tagset: None = ..., - ) -> Incomplete: ... - -class AlpinoCorpusReader(BracketParseCorpusReader): - def __init__( - self, - root: Incomplete, - encoding: str = "ISO-8859-1", - tagset: Incomplete | None = None, - ) -> None: ... diff --git a/stubs/nltk/corpus/reader/categorized_sents.pyi b/stubs/nltk/corpus/reader/categorized_sents.pyi deleted file mode 100644 index 12aaf6d..0000000 --- a/stubs/nltk/corpus/reader/categorized_sents.pyi +++ /dev/null @@ -1,26 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.tokenize import * - -class CategorizedSentencesCorpusReader(CategorizedCorpusReader, CorpusReader): - CorpusView = StreamBackedCorpusView - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - word_tokenizer: Incomplete = ..., - sent_tokenizer: Incomplete | None = None, - encoding: str = "utf8", - **kwargs: Incomplete, - ) -> None: ... - def sents( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def words( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/chasen.pyi b/stubs/nltk/corpus/reader/chasen.pyi deleted file mode 100644 index f1bb3b6..0000000 --- a/stubs/nltk/corpus/reader/chasen.pyi +++ /dev/null @@ -1,41 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader import util as util -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * - -class ChasenCorpusReader(CorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - encoding: str = "utf8", - sent_splitter: Incomplete | None = None, - ) -> None: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_words( - self, fileids: Incomplete | None = None - ) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_sents( - self, fileids: Incomplete | None = None - ) -> Incomplete: ... - def paras(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_paras( - self, fileids: Incomplete | None = None - ) -> Incomplete: ... - -class ChasenCorpusView(StreamBackedCorpusView): - def __init__( - self, - corpus_file: Incomplete, - encoding: Incomplete, - tagged: Incomplete, - group_by_sent: Incomplete, - group_by_para: Incomplete, - sent_splitter: Incomplete | None = None, - ) -> None: ... - def read_block(self, stream: Incomplete) -> Incomplete: ... - -def demo() -> None: ... -def test() -> None: ... diff --git a/stubs/nltk/corpus/reader/childes.pyi b/stubs/nltk/corpus/reader/childes.pyi deleted file mode 100644 index 910d09d..0000000 --- a/stubs/nltk/corpus/reader/childes.pyi +++ /dev/null @@ -1,82 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.util import concat as concat -from nltk.corpus.reader.xmldocs import ( - ElementTree as ElementTree, -) -from nltk.corpus.reader.xmldocs import ( - XMLCorpusReader as XMLCorpusReader, -) -from nltk.util import ( - LazyConcatenation as LazyConcatenation, -) -from nltk.util import ( - LazyMap as LazyMap, -) -from nltk.util import ( - flatten as flatten, -) - -__docformat__: str -NS: str - -class CHILDESCorpusReader(XMLCorpusReader): - def __init__( - self, root: Incomplete, fileids: Incomplete, lazy: bool = True - ) -> None: ... - def words( - self, - fileids: Incomplete | None = None, - speaker: str = "ALL", - stem: bool = False, - relation: bool = False, - strip_space: bool = True, - replace: bool = False, - ) -> Incomplete: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - speaker: str = "ALL", - stem: bool = False, - relation: bool = False, - strip_space: bool = True, - replace: bool = False, - ) -> Incomplete: ... - def sents( - self, - fileids: Incomplete | None = None, - speaker: str = "ALL", - stem: bool = False, - relation: Incomplete | None = None, - strip_space: bool = True, - replace: bool = False, - ) -> Incomplete: ... - def tagged_sents( - self, - fileids: Incomplete | None = None, - speaker: str = "ALL", - stem: bool = False, - relation: Incomplete | None = None, - strip_space: bool = True, - replace: bool = False, - ) -> Incomplete: ... - def corpus(self, fileids: Incomplete | None = None) -> Incomplete: ... - def participants( - self, fileids: Incomplete | None = None - ) -> Incomplete: ... - def age( - self, - fileids: Incomplete | None = None, - speaker: str = "CHI", - month: bool = False, - ) -> Incomplete: ... - def convert_age(self, age_year: Incomplete) -> Incomplete: ... - def MLU( - self, fileids: Incomplete | None = None, speaker: str = "CHI" - ) -> Incomplete: ... - childes_url_base: str - def webview_file( - self, fileid: Incomplete, urlbase: Incomplete | None = None - ) -> None: ... - -def demo(corpus_root: Incomplete | None = None) -> None: ... diff --git a/stubs/nltk/corpus/reader/chunked.pyi b/stubs/nltk/corpus/reader/chunked.pyi deleted file mode 100644 index 5920891..0000000 --- a/stubs/nltk/corpus/reader/chunked.pyi +++ /dev/null @@ -1,73 +0,0 @@ -from _typeshed import Incomplete - -from nltk.chunk import tagstr2tree as tagstr2tree -from nltk.corpus.reader.api import * -from nltk.corpus.reader.bracket_parse import ( - BracketParseCorpusReader as BracketParseCorpusReader, -) -from nltk.corpus.reader.util import * -from nltk.tokenize import * -from nltk.tree import Tree as Tree - -class ChunkedCorpusReader(CorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - extension: str = "", - str2chunktree: Incomplete = ..., - sent_tokenizer: Incomplete = ..., - para_block_reader: Incomplete = ..., - encoding: str = "utf8", - tagset: Incomplete | None = None, - ) -> None: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def paras(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_sents( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_paras( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def chunked_words( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def chunked_sents( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def chunked_paras( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - -class ChunkedCorpusView(StreamBackedCorpusView): - def __init__( - self, - fileid: Incomplete, - encoding: Incomplete, - tagged: Incomplete, - group_by_sent: Incomplete, - group_by_para: Incomplete, - chunked: Incomplete, - str2chunktree: Incomplete, - sent_tokenizer: Incomplete, - para_block_reader: Incomplete, - source_tagset: Incomplete | None = None, - target_tagset: Incomplete | None = None, - ) -> None: ... - def read_block(self, stream: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/cmudict.pyi b/stubs/nltk/corpus/reader/cmudict.pyi deleted file mode 100644 index 07c13b7..0000000 --- a/stubs/nltk/corpus/reader/cmudict.pyi +++ /dev/null @@ -1,12 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.util import Index as Index - -class CMUDictCorpusReader(CorpusReader): - def entries(self) -> Incomplete: ... - def words(self) -> Incomplete: ... - def dict(self) -> Incomplete: ... - -def read_cmudict_block(stream: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/comparative_sents.pyi b/stubs/nltk/corpus/reader/comparative_sents.pyi deleted file mode 100644 index ad79d7a..0000000 --- a/stubs/nltk/corpus/reader/comparative_sents.pyi +++ /dev/null @@ -1,45 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.tokenize import * - -STARS: Incomplete -COMPARISON: Incomplete -CLOSE_COMPARISON: Incomplete -GRAD_COMPARISON: Incomplete -NON_GRAD_COMPARISON: Incomplete -ENTITIES_FEATS: Incomplete -KEYWORD: Incomplete - -class Comparison: - text: Incomplete - comp_type: Incomplete - entity_1: Incomplete - entity_2: Incomplete - feature: Incomplete - keyword: Incomplete - def __init__( - self, - text: Incomplete | None = None, - comp_type: Incomplete | None = None, - entity_1: Incomplete | None = None, - entity_2: Incomplete | None = None, - feature: Incomplete | None = None, - keyword: Incomplete | None = None, - ) -> None: ... - -class ComparativeSentencesCorpusReader(CorpusReader): - CorpusView = StreamBackedCorpusView - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - word_tokenizer: Incomplete = ..., - sent_tokenizer: Incomplete | None = None, - encoding: str = "utf8", - ) -> None: ... - def comparisons(self, fileids: Incomplete | None = None) -> Incomplete: ... - def keywords(self, fileids: Incomplete | None = None) -> Incomplete: ... - def keywords_readme(self) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/conll.pyi b/stubs/nltk/corpus/reader/conll.pyi deleted file mode 100644 index 2daf462..0000000 --- a/stubs/nltk/corpus/reader/conll.pyi +++ /dev/null @@ -1,121 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.tag import map_tag as map_tag -from nltk.tree import Tree as Tree -from nltk.util import ( - LazyConcatenation as LazyConcatenation, -) -from nltk.util import ( - LazyMap as LazyMap, -) - -class ConllCorpusReader(CorpusReader): - WORDS: str - POS: str - TREE: str - CHUNK: str - NE: str - SRL: str - IGNORE: str - COLUMN_TYPES: Incomplete - sep: Incomplete - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - columntypes: Incomplete, - chunk_types: Incomplete | None = None, - root_label: str = "S", - pos_in_tree: bool = False, - srl_includes_roleset: bool = True, - encoding: str = "utf8", - tree_class: Incomplete = ..., - tagset: Incomplete | None = None, - separator: Incomplete | None = None, - ) -> None: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_sents( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def chunked_words( - self, - fileids: Incomplete | None = None, - chunk_types: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def chunked_sents( - self, - fileids: Incomplete | None = None, - chunk_types: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def parsed_sents( - self, - fileids: Incomplete | None = None, - pos_in_tree: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def srl_spans(self, fileids: Incomplete | None = None) -> Incomplete: ... - def srl_instances( - self, - fileids: Incomplete | None = None, - pos_in_tree: Incomplete | None = None, - flatten: bool = True, - ) -> Incomplete: ... - def iob_words( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def iob_sents( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - -class ConllSRLInstance: - verb: Incomplete - verb_head: Incomplete - verb_stem: Incomplete - roleset: Incomplete - arguments: Incomplete - tagged_spans: Incomplete - tree: Incomplete - words: Incomplete - def __init__( - self, - tree: Incomplete, - verb_head: Incomplete, - verb_stem: Incomplete, - roleset: Incomplete, - tagged_spans: Incomplete, - ) -> None: ... - def pprint(self) -> Incomplete: ... - -class ConllSRLInstanceList(list): - tree: Incomplete - def __init__( - self, tree: Incomplete, instances: Incomplete = () - ) -> None: ... - def pprint(self, include_tree: bool = False) -> Incomplete: ... - -class ConllChunkCorpusReader(ConllCorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - chunk_types: Incomplete, - encoding: str = "utf8", - tagset: Incomplete | None = None, - separator: Incomplete | None = None, - ) -> None: ... diff --git a/stubs/nltk/corpus/reader/crubadan.pyi b/stubs/nltk/corpus/reader/crubadan.pyi deleted file mode 100644 index c506f91..0000000 --- a/stubs/nltk/corpus/reader/crubadan.pyi +++ /dev/null @@ -1,18 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader import CorpusReader as CorpusReader -from nltk.data import ZipFilePathPointer as ZipFilePathPointer -from nltk.probability import FreqDist as FreqDist - -class CrubadanCorpusReader(CorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - encoding: str = "utf8", - tagset: Incomplete | None = None, - ) -> None: ... - def lang_freq(self, lang: Incomplete) -> Incomplete: ... - def langs(self) -> Incomplete: ... - def iso_to_crubadan(self, lang: Incomplete) -> Incomplete: ... - def crubadan_to_iso(self, lang: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/dependency.pyi b/stubs/nltk/corpus/reader/dependency.pyi deleted file mode 100644 index 68d8e1d..0000000 --- a/stubs/nltk/corpus/reader/dependency.pyi +++ /dev/null @@ -1,44 +0,0 @@ -from typing import ( - Callable, - List, - Optional, - Tuple, - Union, -) - -from nltk.data import ( - SeekableUnicodeStreamReader, - ZipFilePathPointer, -) -from nltk.parse.dependencygraph import DependencyGraph -from nltk.tokenize.regexp import RegexpTokenizer -from nltk.tokenize.simple import TabTokenizer - -class DependencyCorpusReader: - def __init__( - self, - root: ZipFilePathPointer, - fileids: str, - encoding: List[Tuple[str, str]] = ..., - word_tokenizer: TabTokenizer = ..., - sent_tokenizer: RegexpTokenizer = ..., - para_block_reader: Callable = ..., # type: ignore[type-arg] - ): ... - def parsed_sents( - self, fileids: Optional[str] = ... - ) -> List[DependencyGraph]: ... - def sents(self, fileids: Optional[str] = ...) -> DependencyCorpusView: ... - -class DependencyCorpusView: - def __init__( - self, - corpus_file: ZipFilePathPointer, - tagged: bool, - group_by_sent: bool, - dependencies: bool, - chunk_types: None = ..., - encoding: str = ..., - ): ... - def read_block( - self, stream: SeekableUnicodeStreamReader - ) -> List[Union[str, List[str]]]: ... diff --git a/stubs/nltk/corpus/reader/framenet.pyi b/stubs/nltk/corpus/reader/framenet.pyi deleted file mode 100644 index f92d536..0000000 --- a/stubs/nltk/corpus/reader/framenet.pyi +++ /dev/null @@ -1,141 +0,0 @@ -from collections.abc import Generator -from operator import itemgetter as itemgetter - -from _typeshed import Incomplete - -from nltk.corpus.reader import ( - XMLCorpusReader as XMLCorpusReader, -) -from nltk.corpus.reader import ( - XMLCorpusView as XMLCorpusView, -) -from nltk.util import ( - LazyConcatenation as LazyConcatenation, -) -from nltk.util import ( - LazyIteratorList as LazyIteratorList, -) -from nltk.util import ( - LazyMap as LazyMap, -) - -__docformat__: str - -def mimic_wrap( - lines: Incomplete, wrap_at: int = 65, **kwargs: Incomplete -) -> Generator[Incomplete, None, None]: ... - -class FramenetError(Exception): ... - -class AttrDict(dict): # type: ignore[type-arg] - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def __setattr__(self, name: Incomplete, value: Incomplete) -> None: ... - def __getattr__(self, name: Incomplete) -> Incomplete: ... - def __getitem__(self, name: Incomplete) -> Incomplete: ... - -class SpecialList(list): # type: ignore[type-arg] - def __init__( - self, typ: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - -class Future: - def __init__( - self, loader: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - def __nonzero__(self) -> Incomplete: ... - def __len__(self) -> int: ... - def __setitem__(self, key: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, key: Incomplete) -> Incomplete: ... - def __getattr__(self, key: Incomplete) -> Incomplete: ... - -class PrettyDict(AttrDict): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class PrettyList(list): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class PrettyLazyMap(LazyMap): ... -class PrettyLazyIteratorList(LazyIteratorList): ... - -class PrettyLazyConcatenation(LazyConcatenation): - def __add__(self, other: Incomplete) -> Incomplete: ... - def __radd__(self, other: Incomplete) -> Incomplete: ... - -class FramenetCorpusReader(XMLCorpusReader): - def warnings(self, v: Incomplete) -> None: ... - def __init__(self, root: Incomplete, fileids: Incomplete) -> None: ... - def help(self, attrname: Incomplete | None = None) -> Incomplete: ... - def buildindexes(self) -> None: ... - def doc(self, fn_docid: Incomplete) -> Incomplete: ... - def frame_by_id( - self, fn_fid: Incomplete, ignorekeys: Incomplete = [] - ) -> Incomplete: ... - def frame_by_name( - self, - fn_fname: Incomplete, - ignorekeys: Incomplete = [], - check_cache: bool = True, - ) -> Incomplete: ... - def frame( - self, fn_fid_or_fname: Incomplete, ignorekeys: Incomplete = [] - ) -> Incomplete: ... - def frames_by_lemma(self, pat: Incomplete) -> Incomplete: ... - def lu_basic(self, fn_luid: Incomplete) -> Incomplete: ... - def lu( - self, - fn_luid: Incomplete, - ignorekeys: Incomplete = [], - luName: Incomplete | None = None, - frameID: Incomplete | None = None, - frameName: Incomplete | None = None, - ) -> Incomplete: ... - def propagate_semtypes(self) -> None: ... - def semtype(self, key: Incomplete) -> Incomplete: ... - def semtype_inherits( - self, st: Incomplete, superST: Incomplete - ) -> Incomplete: ... - def frames(self, name: Incomplete | None = None) -> Incomplete: ... - def frame_ids_and_names( - self, name: Incomplete | None = None - ) -> Incomplete: ... - def fes( - self, name: Incomplete | None = None, frame: Incomplete | None = None - ) -> Incomplete: ... - def lus( - self, name: Incomplete | None = None, frame: Incomplete | None = None - ) -> Incomplete: ... - def lu_ids_and_names( - self, name: Incomplete | None = None - ) -> Incomplete: ... - def docs_metadata(self, name: Incomplete | None = None) -> Incomplete: ... - def docs(self, name: Incomplete | None = None) -> Incomplete: ... - def sents( - self, exemplars: bool = True, full_text: bool = True - ) -> Incomplete: ... - def annotations( - self, - luNamePattern: Incomplete | None = None, - exemplars: bool = True, - full_text: bool = True, - ) -> Incomplete: ... - def exemplars( - self, - luNamePattern: Incomplete | None = None, - frame: Incomplete | None = None, - fe: Incomplete | None = None, - fe2: Incomplete | None = None, - ) -> Incomplete: ... - def ft_sents( - self, docNamePattern: Incomplete | None = None - ) -> Incomplete: ... - def frame_relation_types(self) -> Incomplete: ... - def frame_relations( - self, - frame: Incomplete | None = None, - frame2: Incomplete | None = None, - type: Incomplete | None = None, - ) -> Incomplete: ... - def fe_relations(self) -> Incomplete: ... - def semtypes(self) -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/corpus/reader/framenet_typed.pyi b/stubs/nltk/corpus/reader/framenet_typed.pyi deleted file mode 100644 index 527eda8..0000000 --- a/stubs/nltk/corpus/reader/framenet_typed.pyi +++ /dev/null @@ -1,163 +0,0 @@ -from collections.abc import Generator -from operator import itemgetter as itemgetter - -from _typeshed import Incomplete - -from nltk.corpus.reader import XMLCorpusReader as XMLCorpusReader -from nltk.corpus.reader import XMLCorpusView as XMLCorpusView -from nltk.util import LazyConcatenation as LazyConcatenation -from nltk.util import LazyIteratorList as LazyIteratorList -from nltk.util import LazyMap as LazyMap - -__docformat__: str - -def mimic_wrap( - lines: Incomplete, wrap_at: int = 65, **kwargs: Incomplete -) -> Generator[Incomplete, None, None]: ... - -class FramenetError(Exception): ... - -class AttrDict(dict): - def __init__( - self: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - def __setattr__( - self: Incomplete, name: Incomplete, value: Incomplete - ) -> None: ... - def __getattr__(self: Incomplete, name: Incomplete) -> Incomplete: ... - def __getitem__(self: Incomplete, name: Incomplete) -> Incomplete: ... - -class SpecialList(list): - def __init__( - self: Incomplete, - typ: Incomplete, - *args: Incomplete, - **kwargs: Incomplete, - ) -> None: ... - -class Future: - def __init__( - self: Incomplete, - loader: Incomplete, - *args: Incomplete, - **kwargs: Incomplete, - ) -> None: ... - def __nonzero__(self: Incomplete) -> Incomplete: ... - def __len__(self: Incomplete) -> int: ... - def __setitem__( - self: Incomplete, key: Incomplete, value: Incomplete - ) -> None: ... - def __getitem__(self: Incomplete, key: Incomplete) -> Incomplete: ... - def __getattr__(self: Incomplete, key: Incomplete) -> Incomplete: ... - -class PrettyDict(AttrDict): - def __init__( - self: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - -class PrettyList(list): - def __init__( - self: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - -class PrettyLazyMap(LazyMap): ... -class PrettyLazyIteratorList(LazyIteratorList): ... - -class PrettyLazyConcatenation(LazyConcatenation): - def __add__(self: Incomplete, other: Incomplete) -> Incomplete: ... - def __radd__(self: Incomplete, other: Incomplete) -> Incomplete: ... - -class FramenetCorpusReader(XMLCorpusReader): - def warnings(self: Incomplete, v: Incomplete) -> None: ... - def __init__( - self: Incomplete, root: Incomplete, fileids: Incomplete - ) -> None: ... - def help( - self: Incomplete, attrname: Incomplete | None = None - ) -> Incomplete: ... - def buildindexes(self: Incomplete) -> None: ... - def doc(self: Incomplete, fn_docid: Incomplete) -> Incomplete: ... - def frame_by_id( - self: Incomplete, fn_fid: Incomplete, ignorekeys: Incomplete = [] - ) -> Incomplete: ... - def frame_by_name( - self: Incomplete, - fn_fname: Incomplete, - ignorekeys: Incomplete = [], - check_cache: bool = True, - ) -> Incomplete: ... - def frame( - self: Incomplete, - fn_fid_or_fname: Incomplete, - ignorekeys: Incomplete = [], - ) -> Incomplete: ... - def frames_by_lemma(self: Incomplete, pat: Incomplete) -> Incomplete: ... - def lu_basic(self: Incomplete, fn_luid: Incomplete) -> Incomplete: ... - def lu( - self: Incomplete, - fn_luid: Incomplete, - ignorekeys: Incomplete = [], - luName: Incomplete | None = None, - frameID: Incomplete | None = None, - frameName: Incomplete | None = None, - ) -> Incomplete: ... - def propagate_semtypes(self: Incomplete) -> None: ... - def semtype(self: Incomplete, key: Incomplete) -> Incomplete: ... - def semtype_inherits( - self: Incomplete, st: Incomplete, superST: Incomplete - ) -> Incomplete: ... - def frames( - self: Incomplete, name: Incomplete | None = None - ) -> Incomplete: ... - def frame_ids_and_names( - self: Incomplete, name: Incomplete | None = None - ) -> Incomplete: ... - def fes( - self: Incomplete, - name: Incomplete | None = None, - frame: Incomplete | None = None, - ) -> Incomplete: ... - def lus( - self: Incomplete, - name: Incomplete | None = None, - frame: Incomplete | None = None, - ) -> Incomplete: ... - def lu_ids_and_names( - self: Incomplete, name: Incomplete | None = None - ) -> Incomplete: ... - def docs_metadata( - self: Incomplete, name: Incomplete | None = None - ) -> Incomplete: ... - def docs( - self: Incomplete, name: Incomplete | None = None - ) -> Incomplete: ... - def sents( - self: Incomplete, exemplars: bool = True, full_text: bool = True - ) -> Incomplete: ... - def annotations( - self: Incomplete, - luNamePattern: Incomplete | None = None, - exemplars: bool = True, - full_text: bool = True, - ) -> Incomplete: ... - def exemplars( - self: Incomplete, - luNamePattern: Incomplete | None = None, - frame: Incomplete | None = None, - fe: Incomplete | None = None, - fe2: Incomplete | None = None, - ) -> Incomplete: ... - def ft_sents( - self: Incomplete, docNamePattern: Incomplete | None = None - ) -> Incomplete: ... - def frame_relation_types(self: Incomplete) -> Incomplete: ... - def frame_relations( - self: Incomplete, - frame: Incomplete | None = None, - frame2: Incomplete | None = None, - type: Incomplete | None = None, - ) -> Incomplete: ... - def fe_relations(self: Incomplete) -> Incomplete: ... - def semtypes(self: Incomplete) -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/corpus/reader/ieer.pyi b/stubs/nltk/corpus/reader/ieer.pyi deleted file mode 100644 index a6b646f..0000000 --- a/stubs/nltk/corpus/reader/ieer.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * - -titles: Incomplete -documents: Incomplete - -class IEERDocument: - text: Incomplete - docno: Incomplete - doctype: Incomplete - date_time: Incomplete - headline: Incomplete - def __init__( - self, - text: Incomplete, - docno: Incomplete | None = None, - doctype: Incomplete | None = None, - date_time: Incomplete | None = None, - headline: str = "", - ) -> None: ... - -class IEERCorpusReader(CorpusReader): - def docs(self, fileids: Incomplete | None = None) -> Incomplete: ... - def parsed_docs(self, fileids: Incomplete | None = None) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/indian.pyi b/stubs/nltk/corpus/reader/indian.pyi deleted file mode 100644 index 1608828..0000000 --- a/stubs/nltk/corpus/reader/indian.pyi +++ /dev/null @@ -1,31 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.tag import map_tag as map_tag -from nltk.tag import str2tuple as str2tuple - -class IndianCorpusReader(CorpusReader): - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_sents( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - -class IndianCorpusView(StreamBackedCorpusView): - def __init__( - self, - corpus_file: Incomplete, - encoding: Incomplete, - tagged: Incomplete, - group_by_sent: Incomplete, - tag_mapping_function: Incomplete | None = None, - ) -> None: ... - def read_block(self, stream: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/ipipan.pyi b/stubs/nltk/corpus/reader/ipipan.pyi deleted file mode 100644 index 06c7d22..0000000 --- a/stubs/nltk/corpus/reader/ipipan.pyi +++ /dev/null @@ -1,3 +0,0 @@ -from typing import Callable - -def _parse_args(fun: Callable) -> Callable: ... diff --git a/stubs/nltk/corpus/reader/knbc.pyi b/stubs/nltk/corpus/reader/knbc.pyi deleted file mode 100644 index 2fa9acc..0000000 --- a/stubs/nltk/corpus/reader/knbc.pyi +++ /dev/null @@ -1,31 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import ( - CorpusReader as CorpusReader, -) -from nltk.corpus.reader.api import ( - SyntaxCorpusReader as SyntaxCorpusReader, -) -from nltk.corpus.reader.util import ( - FileSystemPathPointer as FileSystemPathPointer, -) -from nltk.corpus.reader.util import ( - find_corpus_fileids as find_corpus_fileids, -) -from nltk.corpus.reader.util import ( - read_blankline_block as read_blankline_block, -) -from nltk.parse import DependencyGraph as DependencyGraph - -class KNBCorpusReader(SyntaxCorpusReader): - morphs2str: Incomplete - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - encoding: str = "utf8", - morphs2str: Incomplete = ..., - ) -> None: ... - -def demo() -> Incomplete: ... -def test() -> None: ... diff --git a/stubs/nltk/corpus/reader/lin.pyi b/stubs/nltk/corpus/reader/lin.pyi deleted file mode 100644 index 8eac57f..0000000 --- a/stubs/nltk/corpus/reader/lin.pyi +++ /dev/null @@ -1,21 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader import CorpusReader as CorpusReader - -class LinThesaurusCorpusReader(CorpusReader): - def __init__(self, root: Incomplete, badscore: float = 0.0) -> None: ... - def similarity( - self, - ngram1: Incomplete, - ngram2: Incomplete, - fileid: Incomplete | None = None, - ) -> Incomplete: ... - def scored_synonyms( - self, ngram: Incomplete, fileid: Incomplete | None = None - ) -> Incomplete: ... - def synonyms( - self, ngram: Incomplete, fileid: Incomplete | None = None - ) -> Incomplete: ... - def __contains__(self, ngram: Incomplete) -> bool: ... - -def demo() -> None: ... diff --git a/stubs/nltk/corpus/reader/markdown.pyi b/stubs/nltk/corpus/reader/markdown.pyi deleted file mode 100644 index 5326bf8..0000000 --- a/stubs/nltk/corpus/reader/markdown.pyi +++ /dev/null @@ -1,157 +0,0 @@ -from typing import NamedTuple - -from _typeshed import Incomplete - -from nltk.corpus.reader.api import ( - CategorizedCorpusReader as CategorizedCorpusReader, -) -from nltk.corpus.reader.plaintext import ( - PlaintextCorpusReader as PlaintextCorpusReader, -) -from nltk.corpus.reader.util import ( - concat as concat, -) -from nltk.corpus.reader.util import ( - read_blankline_block as read_blankline_block, -) -from nltk.tokenize import ( - blankline_tokenize as blankline_tokenize, -) -from nltk.tokenize import ( - sent_tokenize as sent_tokenize, -) -from nltk.tokenize import ( - word_tokenize as word_tokenize, -) - -def comma_separated_string_args(func: Incomplete) -> Incomplete: ... -def read_parse_blankline_block( - stream: Incomplete, parser: Incomplete -) -> Incomplete: ... - -class MarkdownBlock: - content: Incomplete - truncate_at: int - def __init__(self, content: Incomplete) -> None: ... - @property - def raw(self) -> Incomplete: ... - @property - def words(self) -> Incomplete: ... - @property - def sents(self) -> Incomplete: ... - @property - def paras(self) -> Incomplete: ... - -class CodeBlock(MarkdownBlock): - language: Incomplete - def __init__(self, language: Incomplete, *args: Incomplete) -> None: ... - @property - def sents(self) -> Incomplete: ... - @property - def lines(self) -> Incomplete: ... - @property - def paras(self) -> Incomplete: ... - -class MarkdownSection(MarkdownBlock): - heading: Incomplete - level: Incomplete - def __init__( - self, heading: Incomplete, level: Incomplete, *args: Incomplete - ) -> None: ... - -class Image(NamedTuple): - label: Incomplete - src: Incomplete - title: Incomplete - -class Link(NamedTuple): - label: Incomplete - href: Incomplete - title: Incomplete - -class List(NamedTuple): - is_ordered: Incomplete - items: Incomplete - -class MarkdownCorpusReader(PlaintextCorpusReader): - parser: Incomplete - def __init__( - self, - *args: Incomplete, - parser: Incomplete | None = None, - **kwargs: Incomplete, - ) -> None: ... - -class CategorizedMarkdownCorpusReader( - CategorizedCorpusReader, MarkdownCorpusReader -): - def __init__( - self, *args: Incomplete, cat_field: str = "tags", **kwargs: Incomplete - ) -> None: ... - def categories(self, fileids: Incomplete | None = None) -> Incomplete: ... - def fileids(self, categories: Incomplete | None = None) -> Incomplete: ... - def raw( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def words( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def sents( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def paras( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def concatenated_view( - self, reader: Incomplete, fileids: Incomplete, categories: Incomplete - ) -> Incomplete: ... - def metadata_reader(self, stream: Incomplete) -> Incomplete: ... - def metadata( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def blockquote_reader(self, stream: Incomplete) -> Incomplete: ... - def blockquotes( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def code_block_reader(self, stream: Incomplete) -> Incomplete: ... - def code_blocks( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def image_reader(self, stream: Incomplete) -> Incomplete: ... - def images( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def link_reader(self, stream: Incomplete) -> Incomplete: ... - def links( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def list_reader(self, stream: Incomplete) -> Incomplete: ... - def lists( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def section_reader(self, stream: Incomplete) -> Incomplete: ... - def sections( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/mte.pyi b/stubs/nltk/corpus/reader/mte.pyi deleted file mode 100644 index 57dc458..0000000 --- a/stubs/nltk/corpus/reader/mte.pyi +++ /dev/null @@ -1,90 +0,0 @@ -from functools import reduce as reduce - -from _typeshed import Incomplete - -from nltk.corpus.reader import ( - TaggedCorpusReader as TaggedCorpusReader, -) -from nltk.corpus.reader import ( - concat as concat, -) -from nltk.corpus.reader.xmldocs import XMLCorpusView as XMLCorpusView - -def xpath( - root: Incomplete, path: Incomplete, ns: Incomplete -) -> Incomplete: ... - -class MTECorpusView(XMLCorpusView): - def __init__( - self, - fileid: Incomplete, - tagspec: Incomplete, - elt_handler: Incomplete | None = None, - ) -> None: ... - def read_block( - self, - stream: Incomplete, - tagspec: Incomplete | None = None, - elt_handler: Incomplete | None = None, - ) -> Incomplete: ... - -class MTEFileReader: - ns: Incomplete - tag_ns: str - xml_ns: str - word_path: str - sent_path: str - para_path: str - def __init__(self, file_path: Incomplete) -> None: ... - def words(self) -> Incomplete: ... - def sents(self) -> Incomplete: ... - def paras(self) -> Incomplete: ... - def lemma_words(self) -> Incomplete: ... - def tagged_words( - self, tagset: Incomplete, tags: Incomplete - ) -> Incomplete: ... - def lemma_sents(self) -> Incomplete: ... - def tagged_sents( - self, tagset: Incomplete, tags: Incomplete - ) -> Incomplete: ... - def lemma_paras(self) -> Incomplete: ... - def tagged_paras( - self, tagset: Incomplete, tags: Incomplete - ) -> Incomplete: ... - -class MTETagConverter: - mapping_msd_universal: Incomplete - @staticmethod - def msd_to_universal(tag: Incomplete) -> Incomplete: ... - -class MTECorpusReader(TaggedCorpusReader): - def __init__( - self, - root: Incomplete | None = None, - fileids: Incomplete | None = None, - encoding: str = "utf8", - ) -> None: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def paras(self, fileids: Incomplete | None = None) -> Incomplete: ... - def lemma_words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - tagset: str = "msd", - tags: str = "", - ) -> Incomplete: ... - def lemma_sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_sents( - self, - fileids: Incomplete | None = None, - tagset: str = "msd", - tags: str = "", - ) -> Incomplete: ... - def lemma_paras(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_paras( - self, - fileids: Incomplete | None = None, - tagset: str = "msd", - tags: str = "", - ) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/mte_typed.pyi b/stubs/nltk/corpus/reader/mte_typed.pyi deleted file mode 100644 index 334f54e..0000000 --- a/stubs/nltk/corpus/reader/mte_typed.pyi +++ /dev/null @@ -1,100 +0,0 @@ -from functools import reduce as reduce - -from _typeshed import Incomplete - -from nltk.corpus.reader import TaggedCorpusReader as TaggedCorpusReader -from nltk.corpus.reader import concat as concat -from nltk.corpus.reader.xmldocs import XMLCorpusView as XMLCorpusView - -def xpath( - root: Incomplete, path: Incomplete, ns: Incomplete -) -> Incomplete: ... - -class MTECorpusView(XMLCorpusView): - def __init__( - self: Incomplete, - fileid: Incomplete, - tagspec: Incomplete, - elt_handler: Incomplete | None = None, - ) -> None: ... - def read_block( - self: Incomplete, - stream: Incomplete, - tagspec: Incomplete | None = None, - elt_handler: Incomplete | None = None, - ) -> Incomplete: ... - -class MTEFileReader: - ns: Incomplete - tag_ns: str - xml_ns: str - word_path: str - sent_path: str - para_path: str - - def __init__(self: Incomplete, file_path: Incomplete) -> None: ... - def words(self: Incomplete) -> Incomplete: ... - def sents(self: Incomplete) -> Incomplete: ... - def paras(self: Incomplete) -> Incomplete: ... - def lemma_words(self: Incomplete) -> Incomplete: ... - def tagged_words( - self: Incomplete, tagset: Incomplete, tags: Incomplete - ) -> Incomplete: ... - def lemma_sents(self: Incomplete) -> Incomplete: ... - def tagged_sents( - self: Incomplete, tagset: Incomplete, tags: Incomplete - ) -> Incomplete: ... - def lemma_paras(self: Incomplete) -> Incomplete: ... - def tagged_paras( - self: Incomplete, tagset: Incomplete, tags: Incomplete - ) -> Incomplete: ... - -class MTETagConverter: - mapping_msd_universal: Incomplete - - @staticmethod - def msd_to_universal(tag: Incomplete) -> Incomplete: ... - -class MTECorpusReader(TaggedCorpusReader): - def __init__( - self: Incomplete, - root: Incomplete | None = None, - fileids: Incomplete | None = None, - encoding: str = "utf8", - ) -> None: ... - def words( - self: Incomplete, fileids: Incomplete | None = None - ) -> Incomplete: ... - def sents( - self: Incomplete, fileids: Incomplete | None = None - ) -> Incomplete: ... - def paras( - self: Incomplete, fileids: Incomplete | None = None - ) -> Incomplete: ... - def lemma_words( - self: Incomplete, fileids: Incomplete | None = None - ) -> Incomplete: ... - def tagged_words( - self: Incomplete, - fileids: Incomplete | None = None, - tagset: str = "msd", - tags: str = "", - ) -> Incomplete: ... - def lemma_sents( - self: Incomplete, fileids: Incomplete | None = None - ) -> Incomplete: ... - def tagged_sents( - self: Incomplete, - fileids: Incomplete | None = None, - tagset: str = "msd", - tags: str = "", - ) -> Incomplete: ... - def lemma_paras( - self: Incomplete, fileids: Incomplete | None = None - ) -> Incomplete: ... - def tagged_paras( - self: Incomplete, - fileids: Incomplete | None = None, - tagset: str = "msd", - tags: str = "", - ) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/nkjp.pyi b/stubs/nltk/corpus/reader/nkjp.pyi deleted file mode 100644 index 06c7d22..0000000 --- a/stubs/nltk/corpus/reader/nkjp.pyi +++ /dev/null @@ -1,3 +0,0 @@ -from typing import Callable - -def _parse_args(fun: Callable) -> Callable: ... diff --git a/stubs/nltk/corpus/reader/nombank.pyi b/stubs/nltk/corpus/reader/nombank.pyi deleted file mode 100644 index bbc6b08..0000000 --- a/stubs/nltk/corpus/reader/nombank.pyi +++ /dev/null @@ -1,21 +0,0 @@ -from typing import ( - Callable, - Optional, -) - -from nltk.corpus.reader.util import StreamBackedCorpusView -from nltk.corpus.util import LazyCorpusLoader -from nltk.data import ZipFilePathPointer - -class NombankCorpusReader: - def __init__( - self, - root: ZipFilePathPointer, - nomfile: str, - framefiles: str = ..., - nounsfile: Optional[str] = ..., - parse_fileid_xform: Optional[Callable] = ..., - parse_corpus: Optional[LazyCorpusLoader] = ..., - encoding: str = ..., - ): ... - def nouns(self) -> StreamBackedCorpusView: ... diff --git a/stubs/nltk/corpus/reader/nps_chat.pyi b/stubs/nltk/corpus/reader/nps_chat.pyi deleted file mode 100644 index b30a3ab..0000000 --- a/stubs/nltk/corpus/reader/nps_chat.pyi +++ /dev/null @@ -1,30 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.corpus.reader.xmldocs import * -from nltk.internals import ElementWrapper as ElementWrapper -from nltk.tag import map_tag as map_tag -from nltk.util import LazyConcatenation as LazyConcatenation - -class NPSChatCorpusReader(XMLCorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - wrap_etree: bool = False, - tagset: Incomplete | None = None, - ) -> None: ... - def xml_posts(self, fileids: Incomplete | None = None) -> Incomplete: ... - def posts(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_posts( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/opinion_lexicon.pyi b/stubs/nltk/corpus/reader/opinion_lexicon.pyi deleted file mode 100644 index 9e161c3..0000000 --- a/stubs/nltk/corpus/reader/opinion_lexicon.pyi +++ /dev/null @@ -1,14 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader import WordListCorpusReader as WordListCorpusReader -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import StreamBackedCorpusView - -class IgnoreReadmeCorpusView(StreamBackedCorpusView): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class OpinionLexiconCorpusReader(WordListCorpusReader): - CorpusView = IgnoreReadmeCorpusView - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def positive(self) -> Incomplete: ... - def negative(self) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/panlex_lite.pyi b/stubs/nltk/corpus/reader/panlex_lite.pyi deleted file mode 100644 index 802833f..0000000 --- a/stubs/nltk/corpus/reader/panlex_lite.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import CorpusReader as CorpusReader - -class PanLexLiteCorpusReader(CorpusReader): - MEANING_Q: str - TRANSLATION_Q: str - def __init__(self, root: Incomplete) -> None: ... - def language_varieties( - self, lc: Incomplete | None = None - ) -> Incomplete: ... - def meanings( - self, expr_uid: Incomplete, expr_tt: Incomplete - ) -> Incomplete: ... - def translations( - self, from_uid: Incomplete, from_tt: Incomplete, to_uid: Incomplete - ) -> Incomplete: ... - -class Meaning(dict): - def __init__(self, mn: Incomplete, attr: Incomplete) -> None: ... - def id(self) -> Incomplete: ... - def quality(self) -> Incomplete: ... - def source(self) -> Incomplete: ... - def source_group(self) -> Incomplete: ... - def expressions(self) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/panlex_swadesh.pyi b/stubs/nltk/corpus/reader/panlex_swadesh.pyi deleted file mode 100644 index a42c7b7..0000000 --- a/stubs/nltk/corpus/reader/panlex_swadesh.pyi +++ /dev/null @@ -1,30 +0,0 @@ -from collections.abc import Generator -from typing import NamedTuple - -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.corpus.reader.wordlist import ( - WordListCorpusReader as WordListCorpusReader, -) -from nltk.tokenize import line_tokenize as line_tokenize - -class PanlexLanguage(NamedTuple): - panlex_uid: Incomplete - iso639: Incomplete - iso639_type: Incomplete - script: Incomplete - name: Incomplete - langvar_uid: Incomplete - -class PanlexSwadeshCorpusReader(WordListCorpusReader): - swadesh_size: Incomplete - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def license(self) -> Incomplete: ... - def language_codes(self) -> Incomplete: ... - def get_languages(self) -> Generator[Incomplete, None, None]: ... - def get_macrolanguages(self) -> Incomplete: ... - def words_by_lang(self, lang_code: Incomplete) -> Incomplete: ... - def words_by_iso639(self, iso63_code: Incomplete) -> Incomplete: ... - def entries(self, fileids: Incomplete | None = None) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/pl196x.pyi b/stubs/nltk/corpus/reader/pl196x.pyi deleted file mode 100644 index 8df0654..0000000 --- a/stubs/nltk/corpus/reader/pl196x.pyi +++ /dev/null @@ -1,76 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.xmldocs import XMLCorpusReader as XMLCorpusReader - -PARA: Incomplete -SENT: Incomplete -TAGGEDWORD: Incomplete -WORD: Incomplete -TYPE: Incomplete -ANA: Incomplete -TEXTID: Incomplete - -class TEICorpusView(StreamBackedCorpusView): - def __init__( - self, - corpus_file: Incomplete, - tagged: Incomplete, - group_by_sent: Incomplete, - group_by_para: Incomplete, - tagset: Incomplete | None = None, - head_len: int = 0, - textids: Incomplete | None = None, - ) -> None: ... - def read_block(self, stream: Incomplete) -> Incomplete: ... - -class Pl196xCorpusReader(CategorizedCorpusReader, XMLCorpusReader): - head_len: int - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def decode_tag(self, tag: Incomplete) -> Incomplete: ... - def textids( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def words( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - textids: Incomplete | None = None, - ) -> Incomplete: ... - def sents( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - textids: Incomplete | None = None, - ) -> Incomplete: ... - def paras( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - textids: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - textids: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_sents( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - textids: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_paras( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - textids: Incomplete | None = None, - ) -> Incomplete: ... - def xml( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/plaintext.pyi b/stubs/nltk/corpus/reader/plaintext.pyi deleted file mode 100644 index e932753..0000000 --- a/stubs/nltk/corpus/reader/plaintext.pyi +++ /dev/null @@ -1,34 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.tokenize import * - -class PlaintextCorpusReader(CorpusReader): - CorpusView = StreamBackedCorpusView - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - word_tokenizer: Incomplete = ..., - sent_tokenizer: Incomplete | None = None, - para_block_reader: Incomplete = ..., - encoding: str = "utf8", - ) -> None: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def paras(self, fileids: Incomplete | None = None) -> Incomplete: ... - -class CategorizedPlaintextCorpusReader( - CategorizedCorpusReader, PlaintextCorpusReader -): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class PortugueseCategorizedPlaintextCorpusReader( - CategorizedPlaintextCorpusReader -): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class EuroparlCorpusReader(PlaintextCorpusReader): - def chapters(self, fileids: Incomplete | None = None) -> Incomplete: ... - def paras(self, fileids: Incomplete | None = None) -> None: ... diff --git a/stubs/nltk/corpus/reader/ppattach.pyi b/stubs/nltk/corpus/reader/ppattach.pyi deleted file mode 100644 index 3af7aea..0000000 --- a/stubs/nltk/corpus/reader/ppattach.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * - -class PPAttachment: - sent: Incomplete - verb: Incomplete - noun1: Incomplete - prep: Incomplete - noun2: Incomplete - attachment: Incomplete - def __init__( - self, - sent: Incomplete, - verb: Incomplete, - noun1: Incomplete, - prep: Incomplete, - noun2: Incomplete, - attachment: Incomplete, - ) -> None: ... - -class PPAttachmentCorpusReader(CorpusReader): - def attachments(self, fileids: Incomplete) -> Incomplete: ... - def tuples(self, fileids: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/propbank.pyi b/stubs/nltk/corpus/reader/propbank.pyi deleted file mode 100644 index 240370c..0000000 --- a/stubs/nltk/corpus/reader/propbank.pyi +++ /dev/null @@ -1,115 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.internals import raise_unorderable_types as raise_unorderable_types -from nltk.tree import Tree as Tree - -class PropbankCorpusReader(CorpusReader): - def __init__( - self, - root: Incomplete, - propfile: Incomplete, - framefiles: str = "", - verbsfile: Incomplete | None = None, - parse_fileid_xform: Incomplete | None = None, - parse_corpus: Incomplete | None = None, - encoding: str = "utf8", - ) -> None: ... - def instances(self, baseform: Incomplete | None = None) -> Incomplete: ... - def lines(self) -> Incomplete: ... - def roleset(self, roleset_id: Incomplete) -> Incomplete: ... - def rolesets(self, baseform: Incomplete | None = None) -> Incomplete: ... - def verbs(self) -> Incomplete: ... - -class PropbankInstance: - fileid: Incomplete - sentnum: Incomplete - wordnum: Incomplete - tagger: Incomplete - roleset: Incomplete - inflection: Incomplete - predicate: Incomplete - arguments: Incomplete - parse_corpus: Incomplete - def __init__( - self, - fileid: Incomplete, - sentnum: Incomplete, - wordnum: Incomplete, - tagger: Incomplete, - roleset: Incomplete, - inflection: Incomplete, - predicate: Incomplete, - arguments: Incomplete, - parse_corpus: Incomplete | None = None, - ) -> None: ... - @property - def baseform(self) -> Incomplete: ... - @property - def sensenumber(self) -> Incomplete: ... - @property - def predid(self) -> Incomplete: ... - tree: Incomplete - @staticmethod - def parse( - s: Incomplete, - parse_fileid_xform: Incomplete | None = None, - parse_corpus: Incomplete | None = None, - ) -> Incomplete: ... - -class PropbankPointer: - def __init__(self) -> None: ... - -class PropbankChainTreePointer(PropbankPointer): - pieces: Incomplete - def __init__(self, pieces: Incomplete) -> None: ... - def select(self, tree: Incomplete) -> Incomplete: ... - -class PropbankSplitTreePointer(PropbankPointer): - pieces: Incomplete - def __init__(self, pieces: Incomplete) -> None: ... - def select(self, tree: Incomplete) -> Incomplete: ... - -class PropbankTreePointer(PropbankPointer): - wordnum: Incomplete - height: Incomplete - def __init__(self, wordnum: Incomplete, height: Incomplete) -> None: ... - @staticmethod - def parse(s: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def select(self, tree: Incomplete) -> Incomplete: ... - def treepos(self, tree: Incomplete) -> Incomplete: ... - -class PropbankInflection: - INFINITIVE: str - GERUND: str - PARTICIPLE: str - FINITE: str - FUTURE: str - PAST: str - PRESENT: str - PERFECT: str - PROGRESSIVE: str - PERFECT_AND_PROGRESSIVE: str - THIRD_PERSON: str - ACTIVE: str - PASSIVE: str - NONE: str - form: Incomplete - tense: Incomplete - aspect: Incomplete - person: Incomplete - voice: Incomplete - def __init__( - self, - form: str = "-", - tense: str = "-", - aspect: str = "-", - person: str = "-", - voice: str = "-", - ) -> None: ... - @staticmethod - def parse(s: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/pros_cons.pyi b/stubs/nltk/corpus/reader/pros_cons.pyi deleted file mode 100644 index bc16729..0000000 --- a/stubs/nltk/corpus/reader/pros_cons.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.tokenize import * - -class ProsConsCorpusReader(CategorizedCorpusReader, CorpusReader): - CorpusView = StreamBackedCorpusView - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - word_tokenizer: Incomplete = ..., - encoding: str = "utf8", - **kwargs: Incomplete, - ) -> None: ... - def sents( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... - def words( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - ) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/reviews.pyi b/stubs/nltk/corpus/reader/reviews.pyi deleted file mode 100644 index a9106d1..0000000 --- a/stubs/nltk/corpus/reader/reviews.pyi +++ /dev/null @@ -1,46 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.tokenize import * - -TITLE: Incomplete -FEATURES: Incomplete -NOTES: Incomplete -SENT: Incomplete - -class Review: - title: Incomplete - review_lines: Incomplete - def __init__( - self, - title: Incomplete | None = None, - review_lines: Incomplete | None = None, - ) -> None: ... - def add_line(self, review_line: Incomplete) -> None: ... - def features(self) -> Incomplete: ... - def sents(self) -> Incomplete: ... - -class ReviewLine: - sent: Incomplete - features: Incomplete - notes: Incomplete - def __init__( - self, - sent: Incomplete, - features: Incomplete | None = None, - notes: Incomplete | None = None, - ) -> None: ... - -class ReviewsCorpusReader(CorpusReader): - CorpusView = StreamBackedCorpusView - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - word_tokenizer: Incomplete = ..., - encoding: str = "utf8", - ) -> None: ... - def features(self, fileids: Incomplete | None = None) -> Incomplete: ... - def reviews(self, fileids: Incomplete | None = None) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/rte.pyi b/stubs/nltk/corpus/reader/rte.pyi deleted file mode 100644 index 194c019..0000000 --- a/stubs/nltk/corpus/reader/rte.pyi +++ /dev/null @@ -1,31 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.corpus.reader.xmldocs import * - -def norm(value_strin: Incomplete) -> Incomplete: ... - -class RTEPair: - challenge: Incomplete - id: Incomplete - gid: Incomplete - text: Incomplete - hyp: Incomplete - value: Incomplete - task: Incomplete - length: Incomplete - def __init__( - self, - pair: Incomplete, - challenge: Incomplete | None = None, - id: Incomplete | None = None, - text: Incomplete | None = None, - hyp: Incomplete | None = None, - value: Incomplete | None = None, - task: Incomplete | None = None, - length: Incomplete | None = None, - ) -> None: ... - -class RTECorpusReader(XMLCorpusReader): - def pairs(self, fileids: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/semcor.pyi b/stubs/nltk/corpus/reader/semcor.pyi deleted file mode 100644 index 33afaaa..0000000 --- a/stubs/nltk/corpus/reader/semcor.pyi +++ /dev/null @@ -1,51 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.xmldocs import ( - XMLCorpusReader as XMLCorpusReader, -) -from nltk.corpus.reader.xmldocs import ( - XMLCorpusView as XMLCorpusView, -) -from nltk.tree import Tree as Tree - -__docformat__: str - -class SemcorCorpusReader(XMLCorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - wordnet: Incomplete, - lazy: bool = True, - ) -> None: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def chunks(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_chunks( - self, fileids: Incomplete | None = None, tag: Incomplete = ... - ) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def chunk_sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_sents( - self, fileids: Incomplete | None = None, tag: Incomplete = ... - ) -> Incomplete: ... - -class SemcorSentence(list): - num: Incomplete - def __init__(self, num: Incomplete, items: Incomplete) -> None: ... - -class SemcorWordView(XMLCorpusView): - def __init__( - self, - fileid: Incomplete, - unit: Incomplete, - bracket_sent: Incomplete, - pos_tag: Incomplete, - sem_tag: Incomplete, - wordnet: Incomplete, - ) -> None: ... - def handle_elt( - self, elt: Incomplete, context: Incomplete - ) -> Incomplete: ... - def handle_word(self, elt: Incomplete) -> Incomplete: ... - def handle_sent(self, elt: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/senseval.pyi b/stubs/nltk/corpus/reader/senseval.pyi deleted file mode 100644 index 33cff03..0000000 --- a/stubs/nltk/corpus/reader/senseval.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.tokenize import * - -class SensevalInstance: - word: Incomplete - senses: Incomplete - position: Incomplete - context: Incomplete - def __init__( - self, - word: Incomplete, - position: Incomplete, - context: Incomplete, - senses: Incomplete, - ) -> None: ... - -class SensevalCorpusReader(CorpusReader): - def instances(self, fileids: Incomplete | None = None) -> Incomplete: ... - -class SensevalCorpusView(StreamBackedCorpusView): - def __init__(self, fileid: Incomplete, encoding: Incomplete) -> None: ... - def read_block(self, stream: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/sentiwordnet.pyi b/stubs/nltk/corpus/reader/sentiwordnet.pyi deleted file mode 100644 index f859e61..0000000 --- a/stubs/nltk/corpus/reader/sentiwordnet.pyi +++ /dev/null @@ -1,24 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.corpus.reader import CorpusReader as CorpusReader - -class SentiWordNetCorpusReader(CorpusReader): - def __init__( - self, root: Incomplete, fileids: Incomplete, encoding: str = "utf-8" - ) -> None: ... - def senti_synset(self, *vals: Incomplete) -> Incomplete: ... - def senti_synsets( - self, string: Incomplete, pos: Incomplete | None = None - ) -> Incomplete: ... - def all_senti_synsets(self) -> Generator[Incomplete, None, None]: ... - -class SentiSynset: - synset: Incomplete - def __init__( - self, pos_score: Incomplete, neg_score: Incomplete, synset: Incomplete - ) -> None: ... - def pos_score(self) -> Incomplete: ... - def neg_score(self) -> Incomplete: ... - def obj_score(self) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/sinica_treebank.pyi b/stubs/nltk/corpus/reader/sinica_treebank.pyi deleted file mode 100644 index 922d35e..0000000 --- a/stubs/nltk/corpus/reader/sinica_treebank.pyi +++ /dev/null @@ -1,13 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.tag import map_tag as map_tag -from nltk.tree import sinica_parse as sinica_parse - -IDENTIFIER: Incomplete -APPENDIX: Incomplete -TAGWORD: Incomplete -WORD: Incomplete - -class SinicaTreebankCorpusReader(SyntaxCorpusReader): ... diff --git a/stubs/nltk/corpus/reader/string_category.pyi b/stubs/nltk/corpus/reader/string_category.pyi deleted file mode 100644 index da06e34..0000000 --- a/stubs/nltk/corpus/reader/string_category.pyi +++ /dev/null @@ -1,14 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * - -class StringCategoryCorpusReader(CorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - delimiter: str = " ", - encoding: str = "utf8", - ) -> None: ... - def tuples(self, fileids: Incomplete | None = None) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/switchboard.pyi b/stubs/nltk/corpus/reader/switchboard.pyi deleted file mode 100644 index 112f3ed..0000000 --- a/stubs/nltk/corpus/reader/switchboard.pyi +++ /dev/null @@ -1,24 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.tag import map_tag as map_tag -from nltk.tag import str2tuple as str2tuple - -class SwitchboardTurn(list): - speaker: Incomplete - id: Incomplete - def __init__( - self, words: Incomplete, speaker: Incomplete, id: Incomplete - ) -> None: ... - -class SwitchboardCorpusReader(CorpusReader): - def __init__( - self, root: Incomplete, tagset: Incomplete | None = None - ) -> None: ... - def words(self) -> Incomplete: ... - def tagged_words(self, tagset: Incomplete | None = None) -> Incomplete: ... - def turns(self) -> Incomplete: ... - def tagged_turns(self, tagset: Incomplete | None = None) -> Incomplete: ... - def discourses(self) -> Incomplete: ... - def tagged_discourses(self, tagset: bool = False) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/tagged.pyi b/stubs/nltk/corpus/reader/tagged.pyi deleted file mode 100644 index 9c37555..0000000 --- a/stubs/nltk/corpus/reader/tagged.pyi +++ /dev/null @@ -1,92 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.timit import read_timit_block as read_timit_block -from nltk.corpus.reader.util import * -from nltk.tag import map_tag as map_tag -from nltk.tag import str2tuple as str2tuple -from nltk.tokenize import * - -class TaggedCorpusReader(CorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - sep: str = "/", - word_tokenizer: Incomplete = ..., - sent_tokenizer: Incomplete = ..., - para_block_reader: Incomplete = ..., - encoding: str = "utf8", - tagset: Incomplete | None = None, - ) -> None: ... - def words(self, fileids: Incomplete | None = None) -> Incomplete: ... - def sents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def paras(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_sents( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_paras( - self, - fileids: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - -class CategorizedTaggedCorpusReader( - CategorizedCorpusReader, TaggedCorpusReader -): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def tagged_words( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_sents( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - def tagged_paras( - self, - fileids: Incomplete | None = None, - categories: Incomplete | None = None, - tagset: Incomplete | None = None, - ) -> Incomplete: ... - -class TaggedCorpusView(StreamBackedCorpusView): - def __init__( - self, - corpus_file: Incomplete, - encoding: Incomplete, - tagged: Incomplete, - group_by_sent: Incomplete, - group_by_para: Incomplete, - sep: Incomplete, - word_tokenizer: Incomplete, - sent_tokenizer: Incomplete, - para_block_reader: Incomplete, - tag_mapping_function: Incomplete | None = None, - ) -> None: ... - def read_block(self, stream: Incomplete) -> Incomplete: ... - -class MacMorphoCorpusReader(TaggedCorpusReader): - def __init__( - self, - root: Incomplete, - fileids: Incomplete, - encoding: str = "utf8", - tagset: Incomplete | None = None, - ) -> None: ... - -class TimitTaggedCorpusReader(TaggedCorpusReader): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def paras(self) -> None: ... - def tagged_paras(self) -> None: ... diff --git a/stubs/nltk/corpus/reader/timit.pyi b/stubs/nltk/corpus/reader/timit.pyi deleted file mode 100644 index 9fbbb3c..0000000 --- a/stubs/nltk/corpus/reader/timit.pyi +++ /dev/null @@ -1,86 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.internals import import_from_stdlib as import_from_stdlib -from nltk.tree import Tree as Tree - -class TimitCorpusReader(CorpusReader): - speakers: Incomplete - def __init__(self, root: Incomplete, encoding: str = "utf8") -> None: ... - def fileids(self, filetype: Incomplete | None = None) -> Incomplete: ... - def utteranceids( - self, - dialect: Incomplete | None = None, - sex: Incomplete | None = None, - spkrid: Incomplete | None = None, - sent_type: Incomplete | None = None, - sentid: Incomplete | None = None, - ) -> Incomplete: ... - def transcription_dict(self) -> Incomplete: ... - def spkrid(self, utterance: Incomplete) -> Incomplete: ... - def sentid(self, utterance: Incomplete) -> Incomplete: ... - def utterance( - self, spkrid: Incomplete, sentid: Incomplete - ) -> Incomplete: ... - def spkrutteranceids(self, speaker: Incomplete) -> Incomplete: ... - def spkrinfo(self, speaker: Incomplete) -> Incomplete: ... - def phones(self, utterances: Incomplete | None = None) -> Incomplete: ... - def phone_times( - self, utterances: Incomplete | None = None - ) -> Incomplete: ... - def words(self, utterances: Incomplete | None = None) -> Incomplete: ... - def word_times( - self, utterances: Incomplete | None = None - ) -> Incomplete: ... - def sents(self, utterances: Incomplete | None = None) -> Incomplete: ... - def sent_times( - self, utterances: Incomplete | None = None - ) -> Incomplete: ... - def phone_trees( - self, utterances: Incomplete | None = None - ) -> Incomplete: ... - def wav( - self, - utterance: Incomplete, - start: int = 0, - end: Incomplete | None = None, - ) -> Incomplete: ... - def audiodata( - self, - utterance: Incomplete, - start: int = 0, - end: Incomplete | None = None, - ) -> Incomplete: ... - def play( - self, - utterance: Incomplete, - start: int = 0, - end: Incomplete | None = None, - ) -> None: ... - -class SpeakerInfo: - id: Incomplete - sex: Incomplete - dr: Incomplete - use: Incomplete - recdate: Incomplete - birthdate: Incomplete - ht: Incomplete - race: Incomplete - edu: Incomplete - comments: Incomplete - def __init__( - self, - id: Incomplete, - sex: Incomplete, - dr: Incomplete, - use: Incomplete, - recdate: Incomplete, - birthdate: Incomplete, - ht: Incomplete, - race: Incomplete, - edu: Incomplete, - comments: Incomplete | None = None, - ) -> None: ... - -def read_timit_block(stream: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/toolbox.pyi b/stubs/nltk/corpus/reader/toolbox.pyi deleted file mode 100644 index cbff599..0000000 --- a/stubs/nltk/corpus/reader/toolbox.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.toolbox import ToolboxData as ToolboxData - -class ToolboxCorpusReader(CorpusReader): - def xml( - self, fileids: Incomplete, key: Incomplete | None = None - ) -> Incomplete: ... - def fields( - self, - fileids: Incomplete, - strip: bool = True, - unwrap: bool = True, - encoding: str = "utf8", - errors: str = "strict", - unicode_fields: Incomplete | None = None, - ) -> Incomplete: ... - def entries( - self, fileids: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def words(self, fileids: Incomplete, key: str = "lx") -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/corpus/reader/twitter.pyi b/stubs/nltk/corpus/reader/twitter.pyi deleted file mode 100644 index 9ad3567..0000000 --- a/stubs/nltk/corpus/reader/twitter.pyi +++ /dev/null @@ -1,26 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import CorpusReader as CorpusReader -from nltk.corpus.reader.util import ( - StreamBackedCorpusView as StreamBackedCorpusView, -) -from nltk.corpus.reader.util import ( - ZipFilePathPointer as ZipFilePathPointer, -) -from nltk.corpus.reader.util import ( - concat as concat, -) -from nltk.tokenize import TweetTokenizer as TweetTokenizer - -class TwitterCorpusReader(CorpusReader): - CorpusView = StreamBackedCorpusView - def __init__( - self, - root: Incomplete, - fileids: Incomplete | None = None, - word_tokenizer: Incomplete = ..., - encoding: str = "utf8", - ) -> None: ... - def docs(self, fileids: Incomplete | None = None) -> Incomplete: ... - def strings(self, fileids: Incomplete | None = None) -> Incomplete: ... - def tokenized(self, fileids: Incomplete | None = None) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/udhr.pyi b/stubs/nltk/corpus/reader/udhr.pyi deleted file mode 100644 index 4e74eeb..0000000 --- a/stubs/nltk/corpus/reader/udhr.pyi +++ /dev/null @@ -1,11 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.plaintext import ( - PlaintextCorpusReader as PlaintextCorpusReader, -) -from nltk.corpus.reader.util import find_corpus_fileids as find_corpus_fileids - -class UdhrCorpusReader(PlaintextCorpusReader): - ENCODINGS: Incomplete - SKIP: Incomplete - def __init__(self, root: str = "udhr") -> None: ... diff --git a/stubs/nltk/corpus/reader/util.pyi b/stubs/nltk/corpus/reader/util.pyi index ad87ea7..308d513 100644 --- a/stubs/nltk/corpus/reader/util.pyi +++ b/stubs/nltk/corpus/reader/util.pyi @@ -1,48 +1,15 @@ -from typing import ( - Any, - Callable, - Iterator, - List, - Optional, - Tuple, - Union, -) +from collections.abc import Callable, Iterator from _typeshed import Incomplete -from nltk.corpus.reader.dependency import DependencyCorpusView -from nltk.data import ( - FileSystemPathPointer, - SeekableUnicodeStreamReader, - ZipFilePathPointer, -) - -def _path_from(parent: str, child: str) -> List[Any]: ... -def concat( - docs: List[Union[StreamBackedCorpusView, DependencyCorpusView, Any]], -) -> Union[ - DependencyCorpusView, ConcatenatedCorpusView, StreamBackedCorpusView -]: ... -def find_corpus_fileids( - root: Union[ZipFilePathPointer, FileSystemPathPointer], regexp: str -) -> List[Union[Any, str]]: ... -def read_blankline_block(stream: SeekableUnicodeStreamReader) -> List[str]: ... -def read_regexp_block( - stream: SeekableUnicodeStreamReader, start_re: str, end_re: None = ... -) -> List[str]: ... - -class ConcatenatedCorpusView: - def __init__(self, corpus_views: List[StreamBackedCorpusView]) -> None: ... - def iterate_from( - self, start_tok: int - ) -> Iterator[Union[str, List[Tuple[str, str]]]]: ... +from nltk.data import FileSystemPathPointer, ZipFilePathPointer class StreamBackedCorpusView: - def __getitem__(self, i: Union[int, slice]) -> Union[str, List[str]]: ... + def __getitem__(self, i: int | slice) -> str | list[str]: ... def __init__( self, - fileid: Union[ZipFilePathPointer, FileSystemPathPointer], - block_reader: Optional[Callable] = ..., # type: ignore[type-arg] + fileid: ZipFilePathPointer | FileSystemPathPointer, + block_reader: Callable | None = ..., startpos: int = ..., encoding: str = ..., ) -> None: ... diff --git a/stubs/nltk/corpus/reader/verbnet.pyi b/stubs/nltk/corpus/reader/verbnet.pyi deleted file mode 100644 index f09d2ec..0000000 --- a/stubs/nltk/corpus/reader/verbnet.pyi +++ /dev/null @@ -1,37 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.xmldocs import XMLCorpusReader as XMLCorpusReader - -class VerbnetCorpusReader(XMLCorpusReader): - def __init__( - self, root: Incomplete, fileids: Incomplete, wrap_etree: bool = False - ) -> None: ... - def lemmas(self, vnclass: Incomplete | None = None) -> Incomplete: ... - def wordnetids(self, vnclass: Incomplete | None = None) -> Incomplete: ... - def classids( - self, - lemma: Incomplete | None = None, - wordnetid: Incomplete | None = None, - fileid: Incomplete | None = None, - classid: Incomplete | None = None, - ) -> Incomplete: ... - def vnclass(self, fileid_or_classid: Incomplete) -> Incomplete: ... - def fileids(self, vnclass_ids: Incomplete | None = None) -> Incomplete: ... - def frames(self, vnclass: Incomplete) -> Incomplete: ... - def subclasses(self, vnclass: Incomplete) -> Incomplete: ... - def themroles(self, vnclass: Incomplete) -> Incomplete: ... - def longid(self, shortid: Incomplete) -> Incomplete: ... - def shortid(self, longid: Incomplete) -> Incomplete: ... - def pprint(self, vnclass: Incomplete) -> Incomplete: ... - def pprint_subclasses( - self, vnclass: Incomplete, indent: str = "" - ) -> Incomplete: ... - def pprint_members( - self, vnclass: Incomplete, indent: str = "" - ) -> Incomplete: ... - def pprint_themroles( - self, vnclass: Incomplete, indent: str = "" - ) -> Incomplete: ... - def pprint_frames( - self, vnclass: Incomplete, indent: str = "" - ) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/wordlist.pyi b/stubs/nltk/corpus/reader/wordlist.pyi deleted file mode 100644 index a7c77fa..0000000 --- a/stubs/nltk/corpus/reader/wordlist.pyi +++ /dev/null @@ -1,36 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.util import * -from nltk.tokenize import line_tokenize as line_tokenize - -class WordListCorpusReader(CorpusReader): - def words( - self, - fileids: Incomplete | None = None, - ignore_lines_startswith: str = "\n", - ) -> Incomplete: ... - -class SwadeshCorpusReader(WordListCorpusReader): - def entries(self, fileids: Incomplete | None = None) -> Incomplete: ... - -class NonbreakingPrefixesCorpusReader(WordListCorpusReader): - available_langs: Incomplete - def words( - self, - lang: Incomplete | None = None, - fileids: Incomplete | None = None, - ignore_lines_startswith: str = "#", - ) -> Incomplete: ... - -class UnicharsCorpusReader(WordListCorpusReader): - available_categories: Incomplete - def chars( - self, - category: Incomplete | None = None, - fileids: Incomplete | None = None, - ) -> Incomplete: ... - -class MWAPPDBCorpusReader(WordListCorpusReader): - mwa_ppdb_xxxl_file: str - def entries(self, fileids: Incomplete = ...) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/wordnet.pyi b/stubs/nltk/corpus/reader/wordnet.pyi index d3128e7..d9680e2 100644 --- a/stubs/nltk/corpus/reader/wordnet.pyi +++ b/stubs/nltk/corpus/reader/wordnet.pyi @@ -1,19 +1,9 @@ -from typing import ( - Any, - Dict, - List, - Optional, - Union, -) +from typing import Any from _typeshed import Incomplete -from nltk.corpus.reader import CorpusReader as CorpusReader from nltk.corpus.util import LazyCorpusLoader -from nltk.data import ( - SeekableUnicodeStreamReader, - ZipFilePathPointer, -) +from nltk.data import SeekableUnicodeStreamReader, ZipFilePathPointer class Lemma: def __init__( @@ -29,7 +19,7 @@ class Lemma: class Synset: def __init__(self, wordnet_corpus_reader: WordNetCorpusReader) -> None: ... - def lemmas(self, lang: str = ...) -> List[Lemma]: ... + def lemmas(self, lang: str = ...) -> list[Lemma]: ... class WordNetCorpusReader: def __init__( @@ -40,15 +30,15 @@ class WordNetCorpusReader: def _load_lemma_pos_offset_map(self) -> Incomplete: ... def _morphy( self, form: str, pos: str, check_exceptions: bool = ... - ) -> List[Union[str, Any]]: ... + ) -> list[str | Any]: ... def _synset_from_pos_and_line( self, pos: str, data_file_line: str ) -> Synset: ... def get_version(self) -> str: ... - def index_sense(self, version: Optional[str] = ...) -> Dict[str, str]: ... - def map_to_many(self, version: str = ...) -> Dict[str, List[str]]: ... - def map_to_one(self, version: str = ...) -> Dict[str, str]: ... - def map_wn(self, version: str = ...) -> Dict[str, str]: ... + def index_sense(self, version: str | None = ...) -> dict[str, str]: ... + def map_to_many(self, version: str = ...) -> dict[str, list[str]]: ... + def map_to_one(self, version: str = ...) -> dict[str, str]: ... + def map_wn(self, version: str = ...) -> dict[str, str]: ... def synset_from_pos_and_offset(self, pos: str, offset: int) -> Synset: ... def synsets( self, @@ -56,8 +46,4 @@ class WordNetCorpusReader: pos: None = ..., lang: str = ..., check_exceptions: bool = ..., - ) -> List[Synset]: ... - -class WordNetICCorpusReader(CorpusReader): - def __init__(self, root: Incomplete, fileids: Incomplete) -> None: ... - def ic(self, icfile: Incomplete) -> Incomplete: ... + ) -> list[Synset]: ... diff --git a/stubs/nltk/corpus/reader/xmldocs.pyi b/stubs/nltk/corpus/reader/xmldocs.pyi deleted file mode 100644 index 27516cd..0000000 --- a/stubs/nltk/corpus/reader/xmldocs.pyi +++ /dev/null @@ -1,33 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import CorpusReader as CorpusReader -from nltk.corpus.reader.util import * -from nltk.data import ( - SeekableUnicodeStreamReader as SeekableUnicodeStreamReader, -) -from nltk.internals import ElementWrapper as ElementWrapper -from nltk.tokenize import WordPunctTokenizer as WordPunctTokenizer - -class XMLCorpusReader(CorpusReader): - def __init__( - self, root: Incomplete, fileids: Incomplete, wrap_etree: bool = False - ) -> None: ... - def xml(self, fileid: Incomplete | None = None) -> Incomplete: ... - def words(self, fileid: Incomplete | None = None) -> Incomplete: ... - -class XMLCorpusView(StreamBackedCorpusView): - def __init__( - self, - fileid: Incomplete, - tagspec: Incomplete, - elt_handler: Incomplete | None = None, - ) -> None: ... - def handle_elt( - self, elt: Incomplete, context: Incomplete - ) -> Incomplete: ... - def read_block( - self, - stream: Incomplete, - tagspec: Incomplete | None = None, - elt_handler: Incomplete | None = None, - ) -> Incomplete: ... diff --git a/stubs/nltk/corpus/reader/ycoe.pyi b/stubs/nltk/corpus/reader/ycoe.pyi deleted file mode 100644 index 171900e..0000000 --- a/stubs/nltk/corpus/reader/ycoe.pyi +++ /dev/null @@ -1,38 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus.reader.api import * -from nltk.corpus.reader.bracket_parse import ( - BracketParseCorpusReader as BracketParseCorpusReader, -) -from nltk.corpus.reader.tagged import TaggedCorpusReader as TaggedCorpusReader -from nltk.corpus.reader.util import * -from nltk.tokenize import RegexpTokenizer as RegexpTokenizer - -class YCOECorpusReader(CorpusReader): - def __init__(self, root: Incomplete, encoding: str = "utf8") -> None: ... - def documents(self, fileids: Incomplete | None = None) -> Incomplete: ... - def fileids(self, documents: Incomplete | None = None) -> Incomplete: ... - def words(self, documents: Incomplete | None = None) -> Incomplete: ... - def sents(self, documents: Incomplete | None = None) -> Incomplete: ... - def paras(self, documents: Incomplete | None = None) -> Incomplete: ... - def tagged_words( - self, documents: Incomplete | None = None - ) -> Incomplete: ... - def tagged_sents( - self, documents: Incomplete | None = None - ) -> Incomplete: ... - def tagged_paras( - self, documents: Incomplete | None = None - ) -> Incomplete: ... - def parsed_sents( - self, documents: Incomplete | None = None - ) -> Incomplete: ... - -class YCOEParseCorpusReader(BracketParseCorpusReader): ... - -class YCOETaggedCorpusReader(TaggedCorpusReader): - def __init__( - self, root: Incomplete, items: Incomplete, encoding: str = "utf8" - ) -> None: ... - -documents: Incomplete diff --git a/stubs/nltk/corpus/util.pyi b/stubs/nltk/corpus/util.pyi index 6c76322..d7efd3d 100644 --- a/stubs/nltk/corpus/util.pyi +++ b/stubs/nltk/corpus/util.pyi @@ -1,14 +1,7 @@ -from typing import ( - Any, - Callable, -) +from typing import Any, Callable from _typeshed import Incomplete -from nltk.corpus.reader.api import CorpusReader - -def _make_bound_method(func: Callable, self: CorpusReader) -> Callable: ... - class LazyCorpusLoader: def __getattr__(self, attr: str) -> Callable: ... def __init__( diff --git a/stubs/nltk/data.pyi b/stubs/nltk/data.pyi index 7ba8c37..d4969d9 100644 --- a/stubs/nltk/data.pyi +++ b/stubs/nltk/data.pyi @@ -1,31 +1,9 @@ import types -import zipfile from abc import ABCMeta, abstractmethod -from gzip import GzipFile from _typeshed import Incomplete -__all__ = [ - "path", - "PathPointer", - "FileSystemPathPointer", - "BufferedGzipFile", - "GzipFileSystemPathPointer", - "GzipFileSystemPathPointer", - "find", - "retrieve", - "FORMATS", - "AUTO_FORMATS", - "load", - "show_cfg", - "clear_cache", - "LazyLoader", - "OpenOnDemandZipFile", - "GzipFileSystemPathPointer", - "SeekableUnicodeStreamReader", -] - -path: Incomplete +path: list[str] class PathPointer(metaclass=ABCMeta): @abstractmethod @@ -35,28 +13,6 @@ class PathPointer(metaclass=ABCMeta): @abstractmethod def join(self, fileid: Incomplete) -> Incomplete: ... -class FileSystemPathPointer(PathPointer, str): - def __init__(self, _path: Incomplete) -> None: ... - @property - def path(self) -> Incomplete: ... - def open(self, encoding: Incomplete | None = None) -> Incomplete: ... - def file_size(self) -> Incomplete: ... - def join(self, fileid: Incomplete) -> Incomplete: ... - -class BufferedGzipFile(GzipFile): - def __init__( - self, - filename: Incomplete | None = None, - mode: Incomplete | None = None, - compresslevel: int = 9, - fileobj: Incomplete | None = None, - **kwargs: Incomplete, - ) -> None: ... - def write(self, data: Incomplete) -> None: ... - -class GzipFileSystemPathPointer(FileSystemPathPointer): - def open(self, encoding: Incomplete | None = None) -> Incomplete: ... - class ZipFilePathPointer(PathPointer): def __init__(self, zipfile: Incomplete, entry: str = "") -> None: ... @property @@ -70,37 +26,6 @@ class ZipFilePathPointer(PathPointer): def find( resource_name: Incomplete, paths: Incomplete | None = None ) -> Incomplete: ... -def retrieve( - resource_url: Incomplete, - filename: Incomplete | None = None, - verbose: bool = True, -) -> None: ... - -FORMATS: Incomplete -AUTO_FORMATS: Incomplete - -def load( - resource_url: Incomplete, - format: str = "auto", - cache: bool = True, - verbose: bool = False, - logic_parser: Incomplete | None = None, - fstruct_reader: Incomplete | None = None, - encoding: Incomplete | None = None, -) -> Incomplete: ... -def show_cfg(resource_url: Incomplete, escape: str = "##") -> None: ... -def clear_cache() -> None: ... - -class LazyLoader: - def __init__(self, _path: Incomplete) -> None: ... - def __getattr__(self, attr: Incomplete) -> Incomplete: ... - -class OpenOnDemandZipFile(zipfile.ZipFile): - def __init__(self, filename: Incomplete) -> None: ... - fp: Incomplete - def read(self, name: Incomplete) -> Incomplete: ... - def write(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def writestr(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... class SeekableUnicodeStreamReader: DEBUG: bool diff --git a/stubs/nltk/decorators.pyi b/stubs/nltk/decorators.pyi deleted file mode 100644 index fdf487f..0000000 --- a/stubs/nltk/decorators.pyi +++ /dev/null @@ -1,6 +0,0 @@ -__all__ = ["decorator", "new_wrapper", "getinfo"] -from _typeshed import Incomplete - -def getinfo(func: Incomplete) -> Incomplete: ... -def new_wrapper(wrapper: Incomplete, model: Incomplete) -> Incomplete: ... -def decorator(caller: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/downloader.pyi b/stubs/nltk/downloader.pyi deleted file mode 100644 index 9df7d04..0000000 --- a/stubs/nltk/downloader.pyi +++ /dev/null @@ -1,225 +0,0 @@ -import threading -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.draw.table import Table as Table -from nltk.draw.util import ShowText as ShowText - -TKINTER: bool -TclError = ValueError - -class Package: - id: Incomplete - name: Incomplete - subdir: Incomplete - url: Incomplete - size: Incomplete - unzipped_size: Incomplete - checksum: Incomplete - svn_revision: Incomplete - copyright: Incomplete - contact: Incomplete - license: Incomplete - author: Incomplete - filename: Incomplete - unzip: Incomplete - def __init__( - self, - id: Incomplete, - url: Incomplete, - name: Incomplete | None = None, - subdir: str = "", - size: Incomplete | None = None, - unzipped_size: Incomplete | None = None, - checksum: Incomplete | None = None, - svn_revision: Incomplete | None = None, - copyright: str = "Unknown", - contact: str = "Unknown", - license: str = "Unknown", - author: str = "Unknown", - unzip: bool = True, - **kw: Incomplete, - ) -> None: ... - @staticmethod - def fromxml(xml: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - -class Collection: - id: Incomplete - name: Incomplete - children: Incomplete - packages: Incomplete - def __init__( - self, - id: Incomplete, - children: Incomplete, - name: Incomplete | None = None, - **kw: Incomplete, - ) -> None: ... - @staticmethod - def fromxml(xml: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - -class DownloaderMessage: ... - -class StartCollectionMessage(DownloaderMessage): - collection: Incomplete - def __init__(self, collection: Incomplete) -> None: ... - -class FinishCollectionMessage(DownloaderMessage): - collection: Incomplete - def __init__(self, collection: Incomplete) -> None: ... - -class StartPackageMessage(DownloaderMessage): - package: Incomplete - def __init__(self, package: Incomplete) -> None: ... - -class FinishPackageMessage(DownloaderMessage): - package: Incomplete - def __init__(self, package: Incomplete) -> None: ... - -class StartDownloadMessage(DownloaderMessage): - package: Incomplete - def __init__(self, package: Incomplete) -> None: ... - -class FinishDownloadMessage(DownloaderMessage): - package: Incomplete - def __init__(self, package: Incomplete) -> None: ... - -class StartUnzipMessage(DownloaderMessage): - package: Incomplete - def __init__(self, package: Incomplete) -> None: ... - -class FinishUnzipMessage(DownloaderMessage): - package: Incomplete - def __init__(self, package: Incomplete) -> None: ... - -class UpToDateMessage(DownloaderMessage): - package: Incomplete - def __init__(self, package: Incomplete) -> None: ... - -class StaleMessage(DownloaderMessage): - package: Incomplete - def __init__(self, package: Incomplete) -> None: ... - -class ErrorMessage(DownloaderMessage): - package: Incomplete - message: Incomplete - def __init__(self, package: Incomplete, message: Incomplete) -> None: ... - -class ProgressMessage(DownloaderMessage): - progress: Incomplete - def __init__(self, progress: Incomplete) -> None: ... - -class SelectDownloadDirMessage(DownloaderMessage): - download_dir: Incomplete - def __init__(self, download_dir: Incomplete) -> None: ... - -class Downloader: - INDEX_TIMEOUT: Incomplete - DEFAULT_URL: str - INSTALLED: str - NOT_INSTALLED: str - STALE: str - PARTIAL: str - def __init__( - self, - server_index_url: Incomplete | None = None, - download_dir: Incomplete | None = None, - ) -> None: ... - def list( - self, - download_dir: Incomplete | None = None, - show_packages: bool = True, - show_collections: bool = True, - header: bool = True, - more_prompt: bool = False, - skip_installed: bool = False, - ) -> None: ... - def packages(self) -> Incomplete: ... - def corpora(self) -> Incomplete: ... - def models(self) -> Incomplete: ... - def collections(self) -> Incomplete: ... - def incr_download( - self, - info_or_id: Incomplete, - download_dir: Incomplete | None = None, - force: bool = False, - ) -> Generator[Incomplete, Incomplete, None]: ... - def download( - self, - info_or_id: Incomplete | None = None, - download_dir: Incomplete | None = None, - quiet: bool = False, - force: bool = False, - prefix: str = "[nltk_data] ", - halt_on_error: bool = True, - raise_on_error: bool = False, - print_error_to: Incomplete = ..., - ) -> Incomplete: ... - def is_stale( - self, info_or_id: Incomplete, download_dir: Incomplete | None = None - ) -> Incomplete: ... - def is_installed( - self, info_or_id: Incomplete, download_dir: Incomplete | None = None - ) -> Incomplete: ... - def clear_status_cache(self, id: Incomplete | None = None) -> None: ... - def status( - self, info_or_id: Incomplete, download_dir: Incomplete | None = None - ) -> Incomplete: ... - def update( - self, quiet: bool = False, prefix: str = "[nltk_data] " - ) -> None: ... - def index(self) -> Incomplete: ... - def info(self, id: Incomplete) -> Incomplete: ... - def xmlinfo(self, id: Incomplete) -> Incomplete: ... - url: Incomplete - def default_download_dir(self) -> Incomplete: ... - download_dir: Incomplete - -class DownloaderShell: - def __init__(self, dataserver: Incomplete) -> None: ... - def run(self) -> None: ... - -class DownloaderGUI: - COLUMNS: Incomplete - COLUMN_WEIGHTS: Incomplete - COLUMN_WIDTHS: Incomplete - DEFAULT_COLUMN_WIDTH: int - INITIAL_COLUMNS: Incomplete - def __init__( - self, dataserver: Incomplete, use_threads: bool = True - ) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - HELP: Incomplete - def help(self, *e: Incomplete) -> None: ... - def about(self, *e: Incomplete) -> None: ... - class _DownloadThread(threading.Thread): - data_server: Incomplete - items: Incomplete - lock: Incomplete - message_queue: Incomplete - abort: Incomplete - def __init__( - self, - data_server: Incomplete, - items: Incomplete, - lock: Incomplete, - message_queue: Incomplete, - abort: Incomplete, - ) -> None: ... - def run(self) -> None: ... - -def md5_hexdigest(file: Incomplete) -> Incomplete: ... -def unzip( - filename: Incomplete, root: Incomplete, verbose: bool = True -) -> None: ... -def build_index(root: Incomplete, base_url: Incomplete) -> Incomplete: ... - -download: Incomplete - -def download_shell() -> None: ... -def download_gui() -> None: ... -def update() -> None: ... diff --git a/stubs/nltk/draw/__init__.pyi b/stubs/nltk/draw/__init__.pyi deleted file mode 100644 index fc5e9bd..0000000 --- a/stubs/nltk/draw/__init__.pyi +++ /dev/null @@ -1,26 +0,0 @@ -from nltk.draw.cfg import ( - CFGDemo as CFGDemo, -) -from nltk.draw.cfg import ( - CFGEditor as CFGEditor, -) -from nltk.draw.cfg import ( - ProductionList as ProductionList, -) -from nltk.draw.dispersion import dispersion_plot as dispersion_plot -from nltk.draw.table import Table as Table -from nltk.draw.tree import ( - TreeSegmentWidget as TreeSegmentWidget, -) -from nltk.draw.tree import ( - TreeView as TreeView, -) -from nltk.draw.tree import ( - TreeWidget as TreeWidget, -) -from nltk.draw.tree import ( - draw_trees as draw_trees, -) -from nltk.draw.tree import ( - tree_to_treesegment as tree_to_treesegment, -) diff --git a/stubs/nltk/draw/cfg.pyi b/stubs/nltk/draw/cfg.pyi deleted file mode 100644 index 8604237..0000000 --- a/stubs/nltk/draw/cfg.pyi +++ /dev/null @@ -1,56 +0,0 @@ -from _typeshed import Incomplete - -from nltk.draw.tree import ( - TreeSegmentWidget as TreeSegmentWidget, -) -from nltk.draw.tree import ( - tree_to_treesegment as tree_to_treesegment, -) -from nltk.draw.util import ( - CanvasFrame as CanvasFrame, -) -from nltk.draw.util import ( - ColorizedList as ColorizedList, -) -from nltk.draw.util import ( - ShowText as ShowText, -) -from nltk.draw.util import ( - SymbolWidget as SymbolWidget, -) -from nltk.draw.util import ( - TextWidget as TextWidget, -) -from nltk.grammar import ( - CFG as CFG, -) -from nltk.grammar import ( - Nonterminal as Nonterminal, -) -from nltk.grammar import ( - nonterminals as nonterminals, -) -from nltk.tree import Tree as Tree - -class ProductionList(ColorizedList): - ARROW: Incomplete - -class CFGEditor: - ARROW: Incomplete - def __init__( - self, - parent: Incomplete, - cfg: Incomplete | None = None, - set_cfg_callback: Incomplete | None = None, - ) -> None: ... - -class CFGDemo: - def __init__(self, grammar: Incomplete, text: Incomplete) -> None: ... - def reset_workspace(self) -> None: ... - def workspace_markprod(self, production: Incomplete) -> None: ... - def destroy(self, *args: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -def demo2() -> None: ... -def demo() -> None: ... -def demo3() -> None: ... diff --git a/stubs/nltk/draw/dispersion.pyi b/stubs/nltk/draw/dispersion.pyi deleted file mode 100644 index 3c81de7..0000000 --- a/stubs/nltk/draw/dispersion.pyi +++ /dev/null @@ -1,8 +0,0 @@ -from _typeshed import Incomplete - -def dispersion_plot( - text: Incomplete, - words: Incomplete, - ignore_case: bool = False, - title: str = "Lexical Dispersion Plot", -) -> Incomplete: ... diff --git a/stubs/nltk/draw/table.pyi b/stubs/nltk/draw/table.pyi deleted file mode 100644 index fe373de..0000000 --- a/stubs/nltk/draw/table.pyi +++ /dev/null @@ -1,192 +0,0 @@ -from tkinter import Frame - -from _typeshed import Incomplete - -class MultiListbox(Frame): - FRAME_CONFIG: Incomplete - LABEL_CONFIG: Incomplete - LISTBOX_CONFIG: Incomplete - def __init__( - self, - master: Incomplete, - columns: Incomplete, - column_weights: Incomplete | None = None, - cnf: Incomplete = {}, - **kw: Incomplete, - ) -> None: ... - @property - def column_names(self) -> Incomplete: ... - @property - def column_labels(self) -> Incomplete: ... - @property - def listboxes(self) -> Incomplete: ... - def select( - self, - index: Incomplete | None = None, - delta: Incomplete | None = None, - see: bool = True, - ) -> None: ... - def configure(self, cnf: Incomplete = {}, **kw: Incomplete) -> None: ... - def __setitem__(self, key: Incomplete, val: Incomplete) -> None: ... - def rowconfigure( - self, row_index: Incomplete, cnf: Incomplete = {}, **kw: Incomplete - ) -> None: ... - def columnconfigure( - self, col_index: Incomplete, cnf: Incomplete = {}, **kw: Incomplete - ) -> None: ... - def itemconfigure( - self, - row_index: Incomplete, - col_index: Incomplete, - cnf: Incomplete | None = None, - **kw: Incomplete, - ) -> Incomplete: ... - def insert(self, index: Incomplete, *rows: Incomplete) -> None: ... - def get( - self, first: Incomplete, last: Incomplete | None = None - ) -> Incomplete: ... - def bbox(self, row: Incomplete, col: Incomplete) -> Incomplete: ... - def hide_column(self, col_index: Incomplete) -> None: ... - def show_column(self, col_index: Incomplete) -> None: ... - def bind_to_labels( - self, - sequence: Incomplete | None = None, - func: Incomplete | None = None, - add: Incomplete | None = None, - ) -> Incomplete: ... - def bind_to_listboxes( - self, - sequence: Incomplete | None = None, - func: Incomplete | None = None, - add: Incomplete | None = None, - ) -> None: ... - def bind_to_columns( - self, - sequence: Incomplete | None = None, - func: Incomplete | None = None, - add: Incomplete | None = None, - ) -> Incomplete: ... - def curselection( - self, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def selection_includes( - self, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def itemcget( - self, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def size(self, *args: Incomplete, **kwargs: Incomplete) -> Incomplete: ... - def index(self, *args: Incomplete, **kwargs: Incomplete) -> Incomplete: ... - def nearest( - self, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def activate(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def delete(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def scan_mark(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def scan_dragto(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def see(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def selection_anchor( - self, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - def selection_clear( - self, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - def selection_set( - self, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - def yview(self, *args: Incomplete, **kwargs: Incomplete) -> Incomplete: ... - def yview_moveto( - self, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - def yview_scroll( - self, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - itemconfig = itemconfigure - rowconfig = rowconfigure - columnconfig = columnconfigure - select_anchor = selection_anchor - select_clear = selection_clear - select_includes = selection_includes - select_set = selection_set - -class Table: - def __init__( - self, - master: Incomplete, - column_names: Incomplete, - rows: Incomplete | None = None, - column_weights: Incomplete | None = None, - scrollbar: bool = True, - click_to_sort: bool = True, - reprfunc: Incomplete | None = None, - cnf: Incomplete = {}, - **kw: Incomplete, - ) -> None: ... - def pack(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def grid(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def focus(self) -> None: ... - def bind( - self, - sequence: Incomplete | None = None, - func: Incomplete | None = None, - add: Incomplete | None = None, - ) -> None: ... - def rowconfigure( - self, row_index: Incomplete, cnf: Incomplete = {}, **kw: Incomplete - ) -> None: ... - def columnconfigure( - self, col_index: Incomplete, cnf: Incomplete = {}, **kw: Incomplete - ) -> None: ... - def itemconfigure( - self, - row_index: Incomplete, - col_index: Incomplete, - cnf: Incomplete | None = None, - **kw: Incomplete, - ) -> Incomplete: ... - def bind_to_labels( - self, - sequence: Incomplete | None = None, - func: Incomplete | None = None, - add: Incomplete | None = None, - ) -> Incomplete: ... - def bind_to_listboxes( - self, - sequence: Incomplete | None = None, - func: Incomplete | None = None, - add: Incomplete | None = None, - ) -> Incomplete: ... - def bind_to_columns( - self, - sequence: Incomplete | None = None, - func: Incomplete | None = None, - add: Incomplete | None = None, - ) -> Incomplete: ... - rowconfig = rowconfigure - columnconfig = columnconfigure - itemconfig = itemconfigure - def insert(self, row_index: Incomplete, rowvalue: Incomplete) -> None: ... - def extend(self, rowvalues: Incomplete) -> None: ... - def append(self, rowvalue: Incomplete) -> None: ... - def clear(self) -> None: ... - def __getitem__(self, index: Incomplete) -> Incomplete: ... - def __setitem__(self, index: Incomplete, val: Incomplete) -> None: ... - def __delitem__(self, row_index: Incomplete) -> None: ... - def __len__(self) -> int: ... - @property - def column_names(self) -> Incomplete: ... - def column_index(self, i: Incomplete) -> Incomplete: ... - def hide_column(self, column_index: Incomplete) -> None: ... - def show_column(self, column_index: Incomplete) -> None: ... - def selected_row(self) -> Incomplete: ... - def select( - self, - index: Incomplete | None = None, - delta: Incomplete | None = None, - see: bool = True, - ) -> None: ... - def sort_by( - self, column_index: Incomplete, order: str = "toggle" - ) -> None: ... - -def demo() -> Incomplete: ... diff --git a/stubs/nltk/draw/tree.pyi b/stubs/nltk/draw/tree.pyi deleted file mode 100644 index d47e666..0000000 --- a/stubs/nltk/draw/tree.pyi +++ /dev/null @@ -1,91 +0,0 @@ -from _typeshed import Incomplete - -from nltk.draw.util import ( - BoxWidget as BoxWidget, -) -from nltk.draw.util import ( - CanvasFrame as CanvasFrame, -) -from nltk.draw.util import ( - CanvasWidget as CanvasWidget, -) -from nltk.draw.util import ( - OvalWidget as OvalWidget, -) -from nltk.draw.util import ( - ParenWidget as ParenWidget, -) -from nltk.draw.util import ( - TextWidget as TextWidget, -) -from nltk.tree import Tree as Tree -from nltk.util import in_idle as in_idle - -class TreeSegmentWidget(CanvasWidget): - def __init__( - self, - canvas: Incomplete, - label: Incomplete, - subtrees: Incomplete, - **attribs: Incomplete, - ) -> None: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - def label(self) -> Incomplete: ... - def subtrees(self) -> Incomplete: ... - def set_label(self, label: Incomplete) -> None: ... - def replace_child( - self, oldchild: Incomplete, newchild: Incomplete - ) -> None: ... - def remove_child(self, child: Incomplete) -> None: ... - def insert_child(self, index: Incomplete, child: Incomplete) -> None: ... - -def tree_to_treesegment( - canvas: Incomplete, - t: Incomplete, - make_node: Incomplete = ..., - make_leaf: Incomplete = ..., - **attribs: Incomplete, -) -> Incomplete: ... - -class TreeWidget(CanvasWidget): - def __init__( - self, - canvas: Incomplete, - t: Incomplete, - make_node: Incomplete = ..., - make_leaf: Incomplete = ..., - **attribs: Incomplete, - ) -> None: ... - def expanded_tree(self, *path_to_tree: Incomplete) -> Incomplete: ... - def collapsed_tree(self, *path_to_tree: Incomplete) -> Incomplete: ... - def bind_click_trees( - self, callback: Incomplete, button: int = 1 - ) -> None: ... - def bind_drag_trees( - self, callback: Incomplete, button: int = 1 - ) -> None: ... - def bind_click_leaves( - self, callback: Incomplete, button: int = 1 - ) -> None: ... - def bind_drag_leaves( - self, callback: Incomplete, button: int = 1 - ) -> None: ... - def bind_click_nodes( - self, callback: Incomplete, button: int = 1 - ) -> None: ... - def bind_drag_nodes( - self, callback: Incomplete, button: int = 1 - ) -> None: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - def toggle_collapsed(self, treeseg: Incomplete) -> None: ... - -class TreeView: - def __init__(self, *trees: Incomplete) -> None: ... - def resize(self, *e: Incomplete) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -def draw_trees(*trees: Incomplete) -> None: ... -def demo() -> Incomplete: ... diff --git a/stubs/nltk/draw/util.pyi b/stubs/nltk/draw/util.pyi deleted file mode 100644 index 3c41b68..0000000 --- a/stubs/nltk/draw/util.pyi +++ /dev/null @@ -1,213 +0,0 @@ -from abc import ABCMeta -from tkinter import Menubutton - -from _typeshed import Incomplete - -from nltk.util import in_idle as in_idle - -class CanvasWidget(metaclass=ABCMeta): - def __init__( - self, - canvas: Incomplete, - parent: Incomplete | None = None, - **attribs: Incomplete, - ) -> None: ... - def bbox(self) -> Incomplete: ... - def width(self) -> Incomplete: ... - def height(self) -> Incomplete: ... - def parent(self) -> Incomplete: ... - def child_widgets(self) -> Incomplete: ... - def canvas(self) -> Incomplete: ... - def move(self, dx: Incomplete, dy: Incomplete) -> None: ... - def moveto( - self, x: Incomplete, y: Incomplete, anchor: str = "NW" - ) -> None: ... - def destroy(self) -> None: ... - def update(self, child: Incomplete) -> None: ... - def manage(self) -> None: ... - def tags(self) -> Incomplete: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - def hide(self) -> None: ... - def show(self) -> None: ... - def hidden(self) -> Incomplete: ... - def bind_click(self, callback: Incomplete, button: int = 1) -> None: ... - def bind_drag(self, callback: Incomplete) -> None: ... - def unbind_click(self, button: int = 1) -> None: ... - def unbind_drag(self) -> None: ... - -class TextWidget(CanvasWidget): - def __init__( - self, canvas: Incomplete, text: Incomplete, **attribs: Incomplete - ) -> None: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - def text(self) -> Incomplete: ... - def set_text(self, text: Incomplete) -> None: ... - -class SymbolWidget(TextWidget): - SYMBOLS: Incomplete - def __init__( - self, canvas: Incomplete, symbol: Incomplete, **attribs: Incomplete - ) -> None: ... - def symbol(self) -> Incomplete: ... - def set_symbol(self, symbol: Incomplete) -> None: ... - @staticmethod - def symbolsheet(size: int = 20) -> None: ... - -class AbstractContainerWidget(CanvasWidget): - def __init__( - self, canvas: Incomplete, child: Incomplete, **attribs: Incomplete - ) -> None: ... - def child(self) -> Incomplete: ... - def set_child(self, child: Incomplete) -> None: ... - -class BoxWidget(AbstractContainerWidget): - def __init__( - self, canvas: Incomplete, child: Incomplete, **attribs: Incomplete - ) -> None: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - -class OvalWidget(AbstractContainerWidget): - def __init__( - self, canvas: Incomplete, child: Incomplete, **attribs: Incomplete - ) -> None: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - RATIO: float - -class ParenWidget(AbstractContainerWidget): - def __init__( - self, canvas: Incomplete, child: Incomplete, **attribs: Incomplete - ) -> None: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - -class BracketWidget(AbstractContainerWidget): - def __init__( - self, canvas: Incomplete, child: Incomplete, **attribs: Incomplete - ) -> None: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - -class SequenceWidget(CanvasWidget): - def __init__( - self, canvas: Incomplete, *children: Incomplete, **attribs: Incomplete - ) -> None: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - children: Incomplete - def replace_child( - self, oldchild: Incomplete, newchild: Incomplete - ) -> None: ... - def remove_child(self, child: Incomplete) -> None: ... - def insert_child(self, index: Incomplete, child: Incomplete) -> None: ... - -class StackWidget(CanvasWidget): - def __init__( - self, canvas: Incomplete, *children: Incomplete, **attribs: Incomplete - ) -> None: ... - def __setitem__(self, attr: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, attr: Incomplete) -> Incomplete: ... - children: Incomplete - def replace_child( - self, oldchild: Incomplete, newchild: Incomplete - ) -> None: ... - def remove_child(self, child: Incomplete) -> None: ... - def insert_child(self, index: Incomplete, child: Incomplete) -> None: ... - -class SpaceWidget(CanvasWidget): - def __init__( - self, - canvas: Incomplete, - width: Incomplete, - height: Incomplete, - **attribs: Incomplete, - ) -> None: ... - def set_width(self, width: Incomplete) -> None: ... - def set_height(self, height: Incomplete) -> None: ... - -class ScrollWatcherWidget(CanvasWidget): - def __init__( - self, canvas: Incomplete, *children: Incomplete, **attribs: Incomplete - ) -> None: ... - def add_child(self, canvaswidget: Incomplete) -> None: ... - def remove_child(self, canvaswidget: Incomplete) -> None: ... - -class CanvasFrame: - def __init__( - self, parent: Incomplete | None = None, **kw: Incomplete - ) -> None: ... - def print_to_file(self, filename: Incomplete | None = None) -> None: ... - def scrollregion(self) -> Incomplete: ... - def canvas(self) -> Incomplete: ... - def add_widget( - self, - canvaswidget: Incomplete, - x: Incomplete | None = None, - y: Incomplete | None = None, - ) -> None: ... - def destroy_widget(self, canvaswidget: Incomplete) -> None: ... - def remove_widget(self, canvaswidget: Incomplete) -> None: ... - def pack(self, cnf: Incomplete = {}, **kw: Incomplete) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class ShowText: - def __init__( - self, - root: Incomplete, - title: Incomplete, - text: Incomplete, - width: Incomplete | None = None, - height: Incomplete | None = None, - **textbox_options: Incomplete, - ) -> None: ... - def find_dimentions( - self, text: Incomplete, width: Incomplete, height: Incomplete - ) -> Incomplete: ... - def destroy(self, *e: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class EntryDialog: - def __init__( - self, - parent: Incomplete, - original_text: str = "", - instructions: str = "", - set_callback: Incomplete | None = None, - title: Incomplete | None = None, - ) -> None: ... - -class ColorizedList: - def __init__( - self, parent: Incomplete, items: Incomplete = [], **options: Incomplete - ) -> None: ... - def get(self, index: Incomplete | None = None) -> Incomplete: ... - def set(self, items: Incomplete) -> None: ... - def unmark(self, item: Incomplete | None = None) -> None: ... - def mark(self, item: Incomplete) -> None: ... - def markonly(self, item: Incomplete) -> None: ... - def view(self, item: Incomplete) -> None: ... - def add_callback(self, event: Incomplete, func: Incomplete) -> None: ... - def remove_callback( - self, event: Incomplete, func: Incomplete | None = None - ) -> None: ... - def pack(self, cnf: Incomplete = {}, **kw: Incomplete) -> None: ... - def grid(self, cnf: Incomplete = {}, **kw: Incomplete) -> None: ... - def focus(self) -> None: ... - -class MutableOptionMenu(Menubutton): - widgetName: str - menuname: Incomplete - def __init__( - self, master: Incomplete, values: Incomplete, **options: Incomplete - ) -> None: ... - def add(self, value: Incomplete) -> None: ... - def set(self, value: Incomplete) -> None: ... - def remove(self, value: Incomplete) -> None: ... - def __getitem__(self, name: Incomplete) -> Incomplete: ... - def destroy(self) -> None: ... - -def demo() -> None: ... diff --git a/stubs/nltk/featstruct.pyi b/stubs/nltk/featstruct.pyi deleted file mode 100644 index 1c80aed..0000000 --- a/stubs/nltk/featstruct.pyi +++ /dev/null @@ -1,277 +0,0 @@ -from _typeshed import Incomplete - -from nltk.sem.logic import SubstituteBindingsI - -__all__ = [ - "FeatStruct", - "FeatDict", - "FeatList", - "unify", - "subsumes", - "conflicts", - "Feature", - "SlashFeature", - "RangeFeature", - "SLASH", - "TYPE", - "FeatStructReader", -] - -class FeatStruct(SubstituteBindingsI): - def __new__( - cls: Incomplete, - features: Incomplete | None = None, - **morefeatures: Incomplete, - ) -> Incomplete: ... - def equal_values( - self, other: Incomplete, check_reentrance: bool = False - ) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - def freeze(self) -> None: ... - def frozen(self) -> Incomplete: ... - def copy(self, deep: bool = True) -> Incomplete: ... - def __deepcopy__(self, memo: Incomplete) -> None: ... - def cyclic(self) -> Incomplete: ... - def walk(self) -> Incomplete: ... - def substitute_bindings(self, bindings: Incomplete) -> Incomplete: ... - def retract_bindings(self, bindings: Incomplete) -> Incomplete: ... - def variables(self) -> Incomplete: ... - def rename_variables( - self, - vars: Incomplete | None = None, - used_vars: Incomplete = (), - new_vars: Incomplete | None = None, - ) -> Incomplete: ... - def remove_variables(self) -> Incomplete: ... - def unify( - self, - other: Incomplete, - bindings: Incomplete | None = None, - trace: bool = False, - fail: Incomplete | None = None, - rename_vars: bool = True, - ) -> Incomplete: ... - def subsumes(self, other: Incomplete) -> Incomplete: ... - -class FeatDict(FeatStruct, dict): # type: ignore[type-arg] - def __init__( - self, features: Incomplete | None = None, **morefeatures: Incomplete - ) -> None: ... - def __getitem__(self, name_or_path: Incomplete) -> Incomplete: ... - def get( - self, name_or_path: Incomplete, default: Incomplete | None = None - ) -> Incomplete: ... - def __contains__(self, name_or_path: Incomplete) -> bool: ... - def has_key(self, name_or_path: Incomplete) -> Incomplete: ... - def __delitem__(self, name_or_path: Incomplete) -> None: ... - def __setitem__( - self, name_or_path: Incomplete, value: Incomplete - ) -> None: ... - clear: Incomplete - pop: Incomplete - popitem: Incomplete - setdefault: Incomplete - def update( - self, features: Incomplete | None = None, **morefeatures: Incomplete - ) -> None: ... - def __deepcopy__(self, memo: Incomplete) -> Incomplete: ... - -class FeatList(FeatStruct, list): # type: ignore[type-arg] - def __init__(self, features: Incomplete = ()) -> None: ... - def __getitem__(self, name_or_path: Incomplete) -> Incomplete: ... - def __delitem__(self, name_or_path: Incomplete) -> None: ... - def __setitem__( - self, name_or_path: Incomplete, value: Incomplete - ) -> None: ... - __iadd__: Incomplete - __imul__: Incomplete - append: Incomplete - extend: Incomplete - insert: Incomplete - pop: Incomplete - remove: Incomplete - reverse: Incomplete - sort: Incomplete - def __deepcopy__(self, memo: Incomplete) -> Incomplete: ... - -class _UnificationFailure: ... - -def unify( - fstruct1: Incomplete, - fstruct2: Incomplete, - bindings: Incomplete | None = None, - trace: bool = False, - fail: Incomplete | None = None, - rename_vars: bool = True, - fs_class: str = "default", -) -> Incomplete: ... - -class _UnificationFailureError(Exception): ... - -def subsumes(fstruct1: Incomplete, fstruct2: Incomplete) -> Incomplete: ... -def conflicts( - fstruct1: Incomplete, fstruct2: Incomplete, trace: int = 0 -) -> Incomplete: ... - -class SubstituteBindingsSequence(SubstituteBindingsI): - def variables(self) -> Incomplete: ... - def substitute_bindings(self, bindings: Incomplete) -> Incomplete: ... - def subst(self, v: Incomplete, bindings: Incomplete) -> Incomplete: ... - -class FeatureValueTuple(SubstituteBindingsSequence, tuple): ... # type: ignore[type-arg] -class FeatureValueSet(SubstituteBindingsSequence, frozenset): ... # type: ignore[type-arg] - -class FeatureValueUnion(SubstituteBindingsSequence, frozenset): # type: ignore[type-arg] - def __new__(cls: Incomplete, values: Incomplete) -> Incomplete: ... - -class FeatureValueConcat(SubstituteBindingsSequence, tuple): # type: ignore[type-arg] - def __new__(cls: Incomplete, values: Incomplete) -> Incomplete: ... - -class Feature: - def __init__( - self, - name: Incomplete, - default: Incomplete | None = None, - display: Incomplete | None = None, - ) -> None: ... - @property - def name(self) -> Incomplete: ... - @property - def default(self) -> Incomplete: ... - @property - def display(self) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - def read_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - parser: Incomplete, - ) -> Incomplete: ... - def unify_base_values( - self, fval1: Incomplete, fval2: Incomplete, bindings: Incomplete - ) -> Incomplete: ... - -class SlashFeature(Feature): - def read_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - parser: Incomplete, - ) -> Incomplete: ... - -class RangeFeature(Feature): - RANGE_RE: Incomplete - def read_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - parser: Incomplete, - ) -> Incomplete: ... - def unify_base_values( - self, fval1: Incomplete, fval2: Incomplete, bindings: Incomplete - ) -> Incomplete: ... - -SLASH: Incomplete -TYPE: Incomplete - -class CustomFeatureValue: - def unify(self, other: Incomplete) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class FeatStructReader: - def __init__( - self, - features: Incomplete = ..., - fdict_class: Incomplete = ..., - flist_class: Incomplete = ..., - logic_parser: Incomplete | None = None, - ) -> None: ... - def fromstring( - self, s: Incomplete, fstruct: Incomplete | None = None - ) -> Incomplete: ... - def read_partial( - self, - s: Incomplete, - position: int = 0, - reentrances: Incomplete | None = None, - fstruct: Incomplete | None = None, - ) -> Incomplete: ... - def read_value( - self, s: Incomplete, position: Incomplete, reentrances: Incomplete - ) -> Incomplete: ... - VALUE_HANDLERS: Incomplete - def read_fstruct_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - match: Incomplete, - ) -> Incomplete: ... - def read_str_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - match: Incomplete, - ) -> Incomplete: ... - def read_int_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - match: Incomplete, - ) -> Incomplete: ... - def read_var_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - match: Incomplete, - ) -> Incomplete: ... - def read_sym_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - match: Incomplete, - ) -> Incomplete: ... - def read_app_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - match: Incomplete, - ) -> Incomplete: ... - def read_logic_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - match: Incomplete, - ) -> Incomplete: ... - def read_tuple_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - match: Incomplete, - ) -> Incomplete: ... - def read_set_value( - self, - s: Incomplete, - position: Incomplete, - reentrances: Incomplete, - match: Incomplete, - ) -> Incomplete: ... diff --git a/stubs/nltk/grammar.pyi b/stubs/nltk/grammar.pyi deleted file mode 100644 index 001c81a..0000000 --- a/stubs/nltk/grammar.pyi +++ /dev/null @@ -1,165 +0,0 @@ -from _typeshed import Incomplete - -from nltk.featstruct import FeatDict -from nltk.probability import ImmutableProbabilisticMixIn - -__all__ = [ - "Nonterminal", - "nonterminals", - "CFG", - "Production", - "PCFG", - "ProbabilisticProduction", - "DependencyGrammar", - "DependencyProduction", - "ProbabilisticDependencyGrammar", - "induce_pcfg", - "read_grammar", -] - -class Nonterminal: - def __init__(self, symbol: Incomplete) -> None: ... - def symbol(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - def __div__(self, rhs: Incomplete) -> Incomplete: ... - def __truediv__(self, rhs: Incomplete) -> Incomplete: ... - -def nonterminals(symbols: Incomplete) -> Incomplete: ... - -class FeatStructNonterminal(FeatDict, Nonterminal): - def __hash__(self) -> Incomplete: ... - def symbol(self) -> Incomplete: ... - -class Production: - def __init__(self, lhs: Incomplete, rhs: Incomplete) -> None: ... - def lhs(self) -> Incomplete: ... - def rhs(self) -> Incomplete: ... - def __len__(self) -> int: ... - def is_nonlexical(self) -> Incomplete: ... - def is_lexical(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class DependencyProduction(Production): ... - -class ProbabilisticProduction(Production, ImmutableProbabilisticMixIn): - def __init__( - self, lhs: Incomplete, rhs: Incomplete, **prob: Incomplete - ) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class CFG: - def __init__( - self, - start: Incomplete, - productions: Incomplete, - calculate_leftcorners: bool = True, - ) -> None: ... - @classmethod - def fromstring( - cls: Incomplete, input: Incomplete, encoding: Incomplete | None = None - ) -> Incomplete: ... - def start(self) -> Incomplete: ... - def productions( - self, - lhs: Incomplete | None = None, - rhs: Incomplete | None = None, - empty: bool = False, - ) -> Incomplete: ... - def leftcorners(self, cat: Incomplete) -> Incomplete: ... - def is_leftcorner( - self, cat: Incomplete, left: Incomplete - ) -> Incomplete: ... - def leftcorner_parents(self, cat: Incomplete) -> Incomplete: ... - def check_coverage(self, tokens: Incomplete) -> None: ... - def is_lexical(self) -> Incomplete: ... - def is_nonlexical(self) -> Incomplete: ... - def min_len(self) -> Incomplete: ... - def max_len(self) -> Incomplete: ... - def is_nonempty(self) -> Incomplete: ... - def is_binarised(self) -> Incomplete: ... - def is_flexible_chomsky_normal_form(self) -> Incomplete: ... - def is_chomsky_normal_form(self) -> Incomplete: ... - def chomsky_normal_form( - self, new_token_padding: str = "@$@", flexible: bool = False - ) -> Incomplete: ... - @classmethod - def remove_unitary_rules( - cls: Incomplete, grammar: Incomplete - ) -> Incomplete: ... - @classmethod - def binarize( - cls: Incomplete, grammar: Incomplete, padding: str = "@$@" - ) -> Incomplete: ... - @classmethod - def eliminate_start( - cls: Incomplete, grammar: Incomplete - ) -> Incomplete: ... - -class FeatureGrammar(CFG): - def __init__(self, start: Incomplete, productions: Incomplete) -> None: ... - @classmethod - def fromstring( - cls: Incomplete, - input: Incomplete, - features: Incomplete | None = None, - logic_parser: Incomplete | None = None, - fstruct_reader: Incomplete | None = None, - encoding: Incomplete | None = None, - ) -> Incomplete: ... - def productions( - self, - lhs: Incomplete | None = None, - rhs: Incomplete | None = None, - empty: bool = False, - ) -> Incomplete: ... - def leftcorners(self, cat: Incomplete) -> None: ... - def leftcorner_parents(self, cat: Incomplete) -> None: ... - -class FeatureValueType: - def __init__(self, value: Incomplete) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class DependencyGrammar: - def __init__(self, productions: Incomplete) -> None: ... - @classmethod - def fromstring(cls: Incomplete, input: Incomplete) -> Incomplete: ... - def contains(self, head: Incomplete, mod: Incomplete) -> Incomplete: ... - def __contains__(self, head_mod: Incomplete) -> bool: ... - -class ProbabilisticDependencyGrammar: - def __init__( - self, productions: Incomplete, events: Incomplete, tags: Incomplete - ) -> None: ... - def contains(self, head: Incomplete, mod: Incomplete) -> Incomplete: ... - -class PCFG(CFG): - EPSILON: float - def __init__( - self, - start: Incomplete, - productions: Incomplete, - calculate_leftcorners: bool = True, - ) -> None: ... - @classmethod - def fromstring( - cls: Incomplete, input: Incomplete, encoding: Incomplete | None = None - ) -> Incomplete: ... - -def induce_pcfg(start: Incomplete, productions: Incomplete) -> Incomplete: ... -def read_grammar( - input: Incomplete, - nonterm_parser: Incomplete, - probabilistic: bool = False, - encoding: Incomplete | None = None, -) -> Incomplete: ... diff --git a/stubs/nltk/help.pyi b/stubs/nltk/help.pyi deleted file mode 100644 index a393189..0000000 --- a/stubs/nltk/help.pyi +++ /dev/null @@ -1,7 +0,0 @@ -from _typeshed import Incomplete - -from nltk.data import find as find - -def brown_tagset(tagpattern: Incomplete | None = None) -> None: ... -def claws5_tagset(tagpattern: Incomplete | None = None) -> None: ... -def upenn_tagset(tagpattern: Incomplete | None = None) -> None: ... diff --git a/stubs/nltk/inference/__init__.pyi b/stubs/nltk/inference/__init__.pyi deleted file mode 100644 index d56c967..0000000 --- a/stubs/nltk/inference/__init__.pyi +++ /dev/null @@ -1,38 +0,0 @@ -from nltk.inference.api import ( - ParallelProverBuilder as ParallelProverBuilder, -) -from nltk.inference.api import ( - ParallelProverBuilderCommand as ParallelProverBuilderCommand, -) -from nltk.inference.discourse import ( - CfgReadingCommand as CfgReadingCommand, -) -from nltk.inference.discourse import ( - DiscourseTester as DiscourseTester, -) -from nltk.inference.discourse import ( - DrtGlueReadingCommand as DrtGlueReadingCommand, -) -from nltk.inference.discourse import ( - ReadingCommand as ReadingCommand, -) -from nltk.inference.mace import Mace as Mace -from nltk.inference.mace import MaceCommand as MaceCommand -from nltk.inference.prover9 import ( - Prover9 as Prover9, -) -from nltk.inference.prover9 import ( - Prover9Command as Prover9Command, -) -from nltk.inference.resolution import ( - ResolutionProver as ResolutionProver, -) -from nltk.inference.resolution import ( - ResolutionProverCommand as ResolutionProverCommand, -) -from nltk.inference.tableau import ( - TableauProver as TableauProver, -) -from nltk.inference.tableau import ( - TableauProverCommand as TableauProverCommand, -) diff --git a/stubs/nltk/inference/api.pyi b/stubs/nltk/inference/api.pyi deleted file mode 100644 index 58895a2..0000000 --- a/stubs/nltk/inference/api.pyi +++ /dev/null @@ -1,143 +0,0 @@ -import threading -from abc import ABCMeta, abstractmethod - -from _typeshed import Incomplete - -class Prover(metaclass=ABCMeta): - def prove( - self, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - verbose: bool = False, - ) -> Incomplete: ... - -class ModelBuilder(metaclass=ABCMeta): - def build_model( - self, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - verbose: bool = False, - ) -> Incomplete: ... - -class TheoremToolCommand(metaclass=ABCMeta): - @abstractmethod - def add_assumptions(self, new_assumptions: Incomplete) -> Incomplete: ... - @abstractmethod - def retract_assumptions( - self, retracted: Incomplete, debug: bool = False - ) -> Incomplete: ... - @abstractmethod - def assumptions(self) -> Incomplete: ... - @abstractmethod - def goal(self) -> Incomplete: ... - @abstractmethod - def print_assumptions(self) -> Incomplete: ... - -class ProverCommand(TheoremToolCommand): - @abstractmethod - def prove(self, verbose: bool = False) -> Incomplete: ... - @abstractmethod - def proof(self, simplify: bool = True) -> Incomplete: ... - @abstractmethod - def get_prover(self) -> Incomplete: ... - -class ModelBuilderCommand(TheoremToolCommand): - @abstractmethod - def build_model(self, verbose: bool = False) -> Incomplete: ... - @abstractmethod - def model(self, format: Incomplete | None = None) -> Incomplete: ... - @abstractmethod - def get_model_builder(self) -> Incomplete: ... - -class BaseTheoremToolCommand(TheoremToolCommand): - def __init__( - self, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - ) -> None: ... - def add_assumptions(self, new_assumptions: Incomplete) -> None: ... - def retract_assumptions( - self, retracted: Incomplete, debug: bool = False - ) -> Incomplete: ... - def assumptions(self) -> Incomplete: ... - def goal(self) -> Incomplete: ... - def print_assumptions(self) -> None: ... - -class BaseProverCommand(BaseTheoremToolCommand, ProverCommand): - def __init__( - self, - prover: Incomplete, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - ) -> None: ... - def prove(self, verbose: bool = False) -> Incomplete: ... - def proof(self, simplify: bool = True) -> Incomplete: ... - def decorate_proof( - self, proof_string: Incomplete, simplify: bool = True - ) -> Incomplete: ... - def get_prover(self) -> Incomplete: ... - -class BaseModelBuilderCommand(BaseTheoremToolCommand, ModelBuilderCommand): - def __init__( - self, - modelbuilder: Incomplete, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - ) -> None: ... - def build_model(self, verbose: bool = False) -> Incomplete: ... - def model(self, format: Incomplete | None = None) -> Incomplete: ... - def get_model_builder(self) -> Incomplete: ... - -class TheoremToolCommandDecorator(TheoremToolCommand): - def __init__(self, command: Incomplete) -> None: ... - def assumptions(self) -> Incomplete: ... - def goal(self) -> Incomplete: ... - def add_assumptions(self, new_assumptions: Incomplete) -> None: ... - def retract_assumptions( - self, retracted: Incomplete, debug: bool = False - ) -> None: ... - def print_assumptions(self) -> None: ... - -class ProverCommandDecorator(TheoremToolCommandDecorator, ProverCommand): - def __init__(self, proverCommand: Incomplete) -> None: ... - def prove(self, verbose: bool = False) -> Incomplete: ... - def proof(self, simplify: bool = True) -> Incomplete: ... - def decorate_proof( - self, proof_string: Incomplete, simplify: bool = True - ) -> Incomplete: ... - def get_prover(self) -> Incomplete: ... - -class ModelBuilderCommandDecorator( - TheoremToolCommandDecorator, ModelBuilderCommand -): - def __init__(self, modelBuilderCommand: Incomplete) -> None: ... - def build_model(self, verbose: bool = False) -> Incomplete: ... - def model(self, format: Incomplete | None = None) -> Incomplete: ... - def get_model_builder(self) -> Incomplete: ... - -class ParallelProverBuilder(Prover, ModelBuilder): - def __init__( - self, prover: Incomplete, modelbuilder: Incomplete - ) -> None: ... - -class ParallelProverBuilderCommand(BaseProverCommand, BaseModelBuilderCommand): - def __init__( - self, - prover: Incomplete, - modelbuilder: Incomplete, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - ) -> None: ... - def prove(self, verbose: bool = False) -> Incomplete: ... - def build_model(self, verbose: bool = False) -> Incomplete: ... - -class TheoremToolThread(threading.Thread): - def __init__( - self, - command: Incomplete, - verbose: Incomplete, - name: Incomplete | None = None, - ) -> None: ... - def run(self) -> None: ... - @property - def result(self) -> Incomplete: ... diff --git a/stubs/nltk/inference/discourse.pyi b/stubs/nltk/inference/discourse.pyi deleted file mode 100644 index a60ec44..0000000 --- a/stubs/nltk/inference/discourse.pyi +++ /dev/null @@ -1,93 +0,0 @@ -from abc import ABCMeta, abstractmethod - -from _typeshed import Incomplete - -from nltk.data import show_cfg as show_cfg -from nltk.inference.mace import MaceCommand as MaceCommand -from nltk.inference.prover9 import Prover9Command as Prover9Command -from nltk.parse import load_parser as load_parser -from nltk.parse.malt import MaltParser as MaltParser -from nltk.sem.drt import ( - AnaphoraResolutionException as AnaphoraResolutionException, -) -from nltk.sem.drt import resolve_anaphora as resolve_anaphora -from nltk.sem.glue import DrtGlue as DrtGlue -from nltk.sem.logic import Expression as Expression -from nltk.tag import RegexpTagger as RegexpTagger - -class ReadingCommand(metaclass=ABCMeta): - @abstractmethod - def parse_to_readings(self, sentence: Incomplete) -> Incomplete: ... - def process_thread(self, sentence_readings: Incomplete) -> Incomplete: ... - @abstractmethod - def combine_readings(self, readings: Incomplete) -> Incomplete: ... - @abstractmethod - def to_fol(self, expression: Incomplete) -> Incomplete: ... - -class CfgReadingCommand(ReadingCommand): - def __init__(self, gramfile: Incomplete | None = None) -> None: ... - def parse_to_readings(self, sentence: Incomplete) -> Incomplete: ... - def combine_readings(self, readings: Incomplete) -> Incomplete: ... - def to_fol(self, expression: Incomplete) -> Incomplete: ... - -class DrtGlueReadingCommand(ReadingCommand): - def __init__( - self, - semtype_file: Incomplete | None = None, - remove_duplicates: bool = False, - depparser: Incomplete | None = None, - ) -> None: ... - def parse_to_readings(self, sentence: Incomplete) -> Incomplete: ... - def process_thread(self, sentence_readings: Incomplete) -> Incomplete: ... - def combine_readings(self, readings: Incomplete) -> Incomplete: ... - def to_fol(self, expression: Incomplete) -> Incomplete: ... - -class DiscourseTester: - def __init__( - self, - input: Incomplete, - reading_command: Incomplete | None = None, - background: Incomplete | None = None, - ) -> None: ... - def sentences(self) -> None: ... - def add_sentence( - self, - sentence: Incomplete, - informchk: bool = False, - consistchk: bool = False, - ) -> None: ... - def retract_sentence( - self, sentence: Incomplete, verbose: bool = True - ) -> None: ... - def grammar(self) -> None: ... - def readings( - self, - sentence: Incomplete | None = None, - threaded: bool = False, - verbose: bool = True, - filter: bool = False, - show_thread_readings: bool = False, - ) -> None: ... - def expand_threads( - self, thread_id: Incomplete, threads: Incomplete | None = None - ) -> Incomplete: ... - def models( - self, - thread_id: Incomplete | None = None, - show: bool = True, - verbose: bool = False, - ) -> None: ... - def add_background( - self, background: Incomplete, verbose: bool = False - ) -> None: ... - def background(self) -> None: ... - @staticmethod - def multiply( - discourse: Incomplete, readings: Incomplete - ) -> Incomplete: ... - -def load_fol(s: Incomplete) -> Incomplete: ... -def discourse_demo(reading_command: Incomplete | None = None) -> None: ... -def drt_discourse_demo(reading_command: Incomplete | None = None) -> None: ... -def spacer(num: int = 30) -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/inference/mace.pyi b/stubs/nltk/inference/mace.pyi deleted file mode 100644 index d3d59f2..0000000 --- a/stubs/nltk/inference/mace.pyi +++ /dev/null @@ -1,42 +0,0 @@ -from _typeshed import Incomplete - -from nltk.inference.api import ( - BaseModelBuilderCommand as BaseModelBuilderCommand, -) -from nltk.inference.api import ( - ModelBuilder as ModelBuilder, -) -from nltk.inference.prover9 import ( - Prover9CommandParent as Prover9CommandParent, -) -from nltk.inference.prover9 import ( - Prover9Parent as Prover9Parent, -) -from nltk.sem import Expression as Expression -from nltk.sem import Valuation as Valuation -from nltk.sem.logic import is_indvar as is_indvar - -class MaceCommand(Prover9CommandParent, BaseModelBuilderCommand): - def __init__( - self, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - max_models: int = 500, - model_builder: Incomplete | None = None, - ) -> None: ... - @property - def valuation(mbc: Incomplete) -> Incomplete: ... - -class Mace(Prover9Parent, ModelBuilder): - def __init__(self, end_size: int = 500) -> None: ... - -def spacer(num: int = 30) -> None: ... -def decode_result(found: Incomplete) -> Incomplete: ... -def test_model_found(arguments: Incomplete) -> None: ... -def test_build_model(arguments: Incomplete) -> None: ... -def test_transform_output(argument_pair: Incomplete) -> None: ... -def test_make_relation_set() -> None: ... - -arguments: Incomplete - -def demo() -> None: ... diff --git a/stubs/nltk/inference/nonmonotonic.pyi b/stubs/nltk/inference/nonmonotonic.pyi deleted file mode 100644 index e0ee666..0000000 --- a/stubs/nltk/inference/nonmonotonic.pyi +++ /dev/null @@ -1,93 +0,0 @@ -from _typeshed import Incomplete - -from nltk.inference.api import ( - Prover as Prover, -) -from nltk.inference.api import ( - ProverCommandDecorator as ProverCommandDecorator, -) -from nltk.inference.prover9 import ( - Prover9 as Prover9, -) -from nltk.inference.prover9 import ( - Prover9Command as Prover9Command, -) -from nltk.sem.logic import ( - AbstractVariableExpression as AbstractVariableExpression, -) -from nltk.sem.logic import ( - AllExpression as AllExpression, -) -from nltk.sem.logic import ( - AndExpression as AndExpression, -) -from nltk.sem.logic import ( - ApplicationExpression as ApplicationExpression, -) -from nltk.sem.logic import ( - BooleanExpression as BooleanExpression, -) -from nltk.sem.logic import ( - EqualityExpression as EqualityExpression, -) -from nltk.sem.logic import ( - ExistsExpression as ExistsExpression, -) -from nltk.sem.logic import ( - Expression as Expression, -) -from nltk.sem.logic import ( - ImpExpression as ImpExpression, -) -from nltk.sem.logic import ( - NegatedExpression as NegatedExpression, -) -from nltk.sem.logic import ( - Variable as Variable, -) -from nltk.sem.logic import ( - VariableExpression as VariableExpression, -) -from nltk.sem.logic import ( - operator as operator, -) -from nltk.sem.logic import ( - unique_variable as unique_variable, -) - -class ProverParseError(Exception): ... - -def get_domain(goal: Incomplete, assumptions: Incomplete) -> Incomplete: ... - -class ClosedDomainProver(ProverCommandDecorator): - def assumptions(self) -> Incomplete: ... - def goal(self) -> Incomplete: ... - def replace_quants( - self, ex: Incomplete, domain: Incomplete - ) -> Incomplete: ... - -class UniqueNamesProver(ProverCommandDecorator): - def assumptions(self) -> Incomplete: ... - -class SetHolder(list): - def __getitem__(self, item: Incomplete) -> Incomplete: ... - -class ClosedWorldProver(ProverCommandDecorator): - def assumptions(self) -> Incomplete: ... - -class PredHolder: - signatures: Incomplete - properties: Incomplete - signature_len: Incomplete - def __init__(self) -> None: ... - def append_sig(self, new_sig: Incomplete) -> None: ... - def append_prop(self, new_prop: Incomplete) -> None: ... - def validate_sig_len(self, new_sig: Incomplete) -> None: ... - -def closed_domain_demo() -> None: ... -def unique_names_demo() -> None: ... -def closed_world_demo() -> None: ... -def combination_prover_demo() -> None: ... -def default_reasoning_demo() -> None: ... -def print_proof(goal: Incomplete, premises: Incomplete) -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/inference/prover9.pyi b/stubs/nltk/inference/prover9.pyi deleted file mode 100644 index 3d36ac8..0000000 --- a/stubs/nltk/inference/prover9.pyi +++ /dev/null @@ -1,87 +0,0 @@ -from _typeshed import Incomplete - -from nltk.inference.api import ( - BaseProverCommand as BaseProverCommand, -) -from nltk.inference.api import ( - Prover as Prover, -) -from nltk.sem.logic import ( - AllExpression as AllExpression, -) -from nltk.sem.logic import ( - AndExpression as AndExpression, -) -from nltk.sem.logic import ( - EqualityExpression as EqualityExpression, -) -from nltk.sem.logic import ( - ExistsExpression as ExistsExpression, -) -from nltk.sem.logic import ( - Expression as Expression, -) -from nltk.sem.logic import ( - IffExpression as IffExpression, -) -from nltk.sem.logic import ( - ImpExpression as ImpExpression, -) -from nltk.sem.logic import ( - NegatedExpression as NegatedExpression, -) -from nltk.sem.logic import ( - OrExpression as OrExpression, -) - -p9_return_codes: Incomplete - -class Prover9CommandParent: - def print_assumptions(self, output_format: str = "nltk") -> None: ... - -class Prover9Command(Prover9CommandParent, BaseProverCommand): - def __init__( - self, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - timeout: int = 60, - prover: Incomplete | None = None, - ) -> None: ... - def decorate_proof( - self, proof_string: Incomplete, simplify: bool = True - ) -> Incomplete: ... - -class Prover9Parent: - def config_prover9( - self, binary_location: Incomplete, verbose: bool = False - ) -> None: ... - def prover9_input( - self, goal: Incomplete, assumptions: Incomplete - ) -> Incomplete: ... - def binary_locations(self) -> Incomplete: ... - -def convert_to_prover9(input: Incomplete) -> Incomplete: ... - -class Prover9(Prover9Parent, Prover): - def __init__(self, timeout: int = 60) -> None: ... - def prover9_input( - self, goal: Incomplete, assumptions: Incomplete - ) -> Incomplete: ... - -class Prover9Exception(Exception): - def __init__( - self, returncode: Incomplete, message: Incomplete - ) -> None: ... - -class Prover9FatalException(Prover9Exception): ... -class Prover9LimitExceededException(Prover9Exception): ... - -def test_config() -> None: ... -def test_convert_to_prover9(expr: Incomplete) -> None: ... -def test_prove(arguments: Incomplete) -> None: ... - -arguments: Incomplete -expressions: Incomplete - -def spacer(num: int = 45) -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/inference/resolution.pyi b/stubs/nltk/inference/resolution.pyi deleted file mode 100644 index 75fb07d..0000000 --- a/stubs/nltk/inference/resolution.pyi +++ /dev/null @@ -1,116 +0,0 @@ -from _typeshed import Incomplete - -from nltk.inference.api import ( - BaseProverCommand as BaseProverCommand, -) -from nltk.inference.api import ( - Prover as Prover, -) -from nltk.sem import skolemize as skolemize -from nltk.sem.logic import ( - AndExpression as AndExpression, -) -from nltk.sem.logic import ( - ApplicationExpression as ApplicationExpression, -) -from nltk.sem.logic import ( - EqualityExpression as EqualityExpression, -) -from nltk.sem.logic import ( - Expression as Expression, -) -from nltk.sem.logic import ( - IndividualVariableExpression as IndividualVariableExpression, -) -from nltk.sem.logic import ( - NegatedExpression as NegatedExpression, -) -from nltk.sem.logic import ( - OrExpression as OrExpression, -) -from nltk.sem.logic import ( - Variable as Variable, -) -from nltk.sem.logic import ( - VariableExpression as VariableExpression, -) -from nltk.sem.logic import ( - is_indvar as is_indvar, -) -from nltk.sem.logic import ( - unique_variable as unique_variable, -) - -class ProverParseError(Exception): ... - -class ResolutionProver(Prover): - ANSWER_KEY: str - -class ResolutionProverCommand(BaseProverCommand): - def __init__( - self, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - prover: Incomplete | None = None, - ) -> None: ... - def prove(self, verbose: bool = False) -> Incomplete: ... - def find_answers(self, verbose: bool = False) -> Incomplete: ... - -class Clause(list): - def __init__(self, data: Incomplete) -> None: ... - def unify( - self, - other: Incomplete, - bindings: Incomplete | None = None, - used: Incomplete | None = None, - skipped: Incomplete | None = None, - debug: bool = False, - ) -> Incomplete: ... - def isSubsetOf(self, other: Incomplete) -> Incomplete: ... - def subsumes(self, other: Incomplete) -> Incomplete: ... - def __getslice__( - self, start: Incomplete, end: Incomplete - ) -> Incomplete: ... - def __sub__(self, other: Incomplete) -> Incomplete: ... - def __add__(self, other: Incomplete) -> Incomplete: ... - def is_tautology(self) -> Incomplete: ... - def free(self) -> Incomplete: ... - def replace( - self, variable: Incomplete, expression: Incomplete - ) -> Incomplete: ... - def substitute_bindings(self, bindings: Incomplete) -> Incomplete: ... - -def clausify(expression: Incomplete) -> Incomplete: ... - -class BindingDict: - d: Incomplete - def __init__(self, binding_list: Incomplete | None = None) -> None: ... - def __setitem__( - self, variable: Incomplete, binding: Incomplete - ) -> None: ... - def __getitem__(self, variable: Incomplete) -> Incomplete: ... - def __contains__(self, item: Incomplete) -> bool: ... - def __add__(self, other: Incomplete) -> Incomplete: ... - def __len__(self) -> int: ... - -def most_general_unification( - a: Incomplete, b: Incomplete, bindings: Incomplete | None = None -) -> Incomplete: ... - -class BindingException(Exception): - def __init__(self, arg: Incomplete) -> None: ... - -class UnificationException(Exception): - def __init__(self, a: Incomplete, b: Incomplete) -> None: ... - -class DebugObject: - enabled: Incomplete - indent: Incomplete - def __init__(self, enabled: bool = True, indent: int = 0) -> None: ... - def __add__(self, i: Incomplete) -> Incomplete: ... - def line(self, line: Incomplete) -> None: ... - -def testResolutionProver() -> None: ... -def resolution_test(e: Incomplete) -> None: ... -def test_clausify() -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/inference/tableau.pyi b/stubs/nltk/inference/tableau.pyi deleted file mode 100644 index f477508..0000000 --- a/stubs/nltk/inference/tableau.pyi +++ /dev/null @@ -1,129 +0,0 @@ -from _typeshed import Incomplete - -from nltk.inference.api import ( - BaseProverCommand as BaseProverCommand, -) -from nltk.inference.api import ( - Prover as Prover, -) -from nltk.internals import Counter as Counter -from nltk.sem.logic import ( - AbstractVariableExpression as AbstractVariableExpression, -) -from nltk.sem.logic import ( - AllExpression as AllExpression, -) -from nltk.sem.logic import ( - AndExpression as AndExpression, -) -from nltk.sem.logic import ( - ApplicationExpression as ApplicationExpression, -) -from nltk.sem.logic import ( - EqualityExpression as EqualityExpression, -) -from nltk.sem.logic import ( - ExistsExpression as ExistsExpression, -) -from nltk.sem.logic import ( - Expression as Expression, -) -from nltk.sem.logic import ( - FunctionVariableExpression as FunctionVariableExpression, -) -from nltk.sem.logic import ( - IffExpression as IffExpression, -) -from nltk.sem.logic import ( - ImpExpression as ImpExpression, -) -from nltk.sem.logic import ( - LambdaExpression as LambdaExpression, -) -from nltk.sem.logic import ( - NegatedExpression as NegatedExpression, -) -from nltk.sem.logic import ( - OrExpression as OrExpression, -) -from nltk.sem.logic import ( - Variable as Variable, -) -from nltk.sem.logic import ( - VariableExpression as VariableExpression, -) -from nltk.sem.logic import ( - unique_variable as unique_variable, -) - -class ProverParseError(Exception): ... - -class TableauProver(Prover): - @staticmethod - def is_atom(e: Incomplete) -> Incomplete: ... - -class TableauProverCommand(BaseProverCommand): - def __init__( - self, - goal: Incomplete | None = None, - assumptions: Incomplete | None = None, - prover: Incomplete | None = None, - ) -> None: ... - -class Agenda: - sets: Incomplete - def __init__(self) -> None: ... - def clone(self) -> Incomplete: ... - def __getitem__(self, index: Incomplete) -> Incomplete: ... - def put( - self, expression: Incomplete, context: Incomplete | None = None - ) -> None: ... - def put_all(self, expressions: Incomplete) -> None: ... - def put_atoms(self, atoms: Incomplete) -> None: ... - def pop_first(self) -> Incomplete: ... - def replace_all(self, old: Incomplete, new: Incomplete) -> None: ... - def mark_alls_fresh(self) -> None: ... - def mark_neqs_fresh(self) -> None: ... - -class Debug: - verbose: Incomplete - indent: Incomplete - lines: Incomplete - def __init__( - self, - verbose: Incomplete, - indent: int = 0, - lines: Incomplete | None = None, - ) -> None: ... - def __add__(self, increment: Incomplete) -> Incomplete: ... - def line(self, data: Incomplete, indent: int = 0) -> None: ... - -class Categories: - ATOM: int - PROP: int - N_ATOM: int - N_PROP: int - APP: int - N_APP: int - N_EQ: int - D_NEG: int - N_ALL: int - N_EXISTS: int - AND: int - N_OR: int - N_IMP: int - OR: int - IMP: int - N_AND: int - IFF: int - N_IFF: int - EQ: int - EXISTS: int - ALL: int - -def testTableauProver() -> None: ... -def testHigherOrderTableauProver() -> None: ... -def tableau_test( - c: Incomplete, ps: Incomplete | None = None, verbose: bool = False -) -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/internals.pyi b/stubs/nltk/internals.pyi deleted file mode 100644 index 9cd0701..0000000 --- a/stubs/nltk/internals.pyi +++ /dev/null @@ -1,187 +0,0 @@ -from abc import ABC -from typing import ( - Callable, - List, - Optional, - Tuple, - Type, - Union, -) - -from _typeshed import Incomplete - -from nltk.corpus.reader.util import ( - ConcatenatedCorpusView, - StreamBackedCorpusView, -) -from nltk.parse.api import ParserI -from nltk.parse.corenlp import ( - CoreNLPDependencyParser, - CoreNLPParser, - GenericCoreNLPParser, -) -from nltk.tag.api import TaggerI -from nltk.tokenize.api import TokenizerI - -def _add_epytext_field( - obj: Callable, field: str, message: str -) -> Incomplete: ... -def _decode_stdoutdata(stdoutdata: bytes) -> str: ... -def _mro( - cls: Union[Type[CoreNLPParser], Type[CoreNLPDependencyParser]], -) -> Union[ - Tuple[ - Type[CoreNLPParser], - Type[GenericCoreNLPParser], - Type[ParserI], - Type[TokenizerI], - Type[ABC], - Type[TaggerI], - Type[object], - ], - Tuple[ - Type[CoreNLPDependencyParser], - Type[GenericCoreNLPParser], - Type[ParserI], - Type[TokenizerI], - Type[ABC], - Type[TaggerI], - Type[object], - ], -]: ... -def deprecated(message: str) -> Callable: ... # type: ignore[type-arg] -def find_binary( - name: str, - path_to_bin: None = ..., - env_vars: List[str] = ..., - searchpath: Tuple[()] = ..., - binary_names: Optional[List[str]] = ..., - url: Optional[str] = ..., - verbose: bool = ..., -) -> Incomplete: ... -def find_binary_iter( - name: str, - path_to_bin: Incomplete | None = None, - env_vars: Incomplete = (), - searchpath: Tuple[()] = (), - binary_names: Optional[List[str]] = None, - url: Optional[str] = None, - verbose: bool = False, -) -> Incomplete: ... -def find_file_iter( - filename: str, - env_vars: List[str] = ..., - searchpath: Tuple[()] = (), - file_names: Optional[List[str]] = ..., - url: Optional[str] = ..., - verbose: bool = False, - finding_dir: bool = False, -) -> Incomplete: ... -def find_jar( - name_pattern: str, - path_to_jar: None = ..., - env_vars: Tuple[str] = ..., - searchpath: Tuple[()] = ..., - url: Optional[str] = ..., - verbose: bool = ..., - is_regex: bool = ..., -) -> Incomplete: ... -def find_jar_iter( - name_pattern: str, - path_to_jar: None = ..., - env_vars: Tuple[str] = ..., - searchpath: Tuple[()] = ..., - url: Optional[str] = ..., - verbose: bool = ..., - is_regex: bool = ..., -) -> Incomplete: ... -def is_writable(path: str) -> Incomplete: ... -def overridden(method: Callable) -> bool: ... -def slice_bounds( - sequence: Union[ConcatenatedCorpusView, StreamBackedCorpusView], - slice_obj: slice, - allow_step: bool = ..., -) -> Tuple[int, int]: ... - -class Counter: - def __init__(self, initial_value: int = 0) -> None: ... - def get(self) -> Incomplete: ... - -def config_java( - bin: Incomplete | None = None, - options: Incomplete | None = None, - verbose: bool = False, -) -> None: ... -def java( - cmd: Incomplete, - classpath: Incomplete | None = None, - stdin: Incomplete | None = None, - stdout: Incomplete | None = None, - stderr: Incomplete | None = None, - blocking: bool = True, -) -> Incomplete: ... - -class ReadError(ValueError): - expected: Incomplete - position: Incomplete - def __init__(self, expected: Incomplete, position: Incomplete) -> None: ... - -def read_str(s: Incomplete, start_position: Incomplete) -> Incomplete: ... -def read_int(s: Incomplete, start_position: Incomplete) -> Incomplete: ... -def read_number(s: Incomplete, start_position: Incomplete) -> Incomplete: ... - -class Deprecated: - def __new__( - cls: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - -def find_file( - filename: Incomplete, - env_vars: Incomplete = (), - searchpath: Incomplete = (), - file_names: Incomplete | None = None, - url: Incomplete | None = None, - verbose: bool = False, -) -> Incomplete: ... -def find_dir( - filename: Incomplete, - env_vars: Incomplete = (), - searchpath: Incomplete = (), - file_names: Incomplete | None = None, - url: Incomplete | None = None, - verbose: bool = False, -) -> Incomplete: ... -def find_jars_within_path(path_to_jars: Incomplete) -> Incomplete: ... -def import_from_stdlib(module: Incomplete) -> Incomplete: ... - -class ElementWrapper: - def __new__(cls: Incomplete, etree: Incomplete) -> Incomplete: ... - def __init__(self, etree: Incomplete) -> None: ... - def unwrap(self) -> Incomplete: ... - def __getattr__(self, attrib: Incomplete) -> Incomplete: ... - def __setattr__( - self, attr: Incomplete, value: Incomplete - ) -> Incomplete: ... - def __delattr__(self, attr: Incomplete) -> Incomplete: ... - def __setitem__(self, index: Incomplete, element: Incomplete) -> None: ... - def __delitem__(self, index: Incomplete) -> None: ... - def __setslice__( - self, start: Incomplete, stop: Incomplete, elements: Incomplete - ) -> None: ... - def __delslice__(self, start: Incomplete, stop: Incomplete) -> None: ... - def __len__(self) -> int: ... - def __getitem__(self, index: Incomplete) -> Incomplete: ... - def __getslice__( - self, start: Incomplete, stop: Incomplete - ) -> Incomplete: ... - def getchildren(self) -> Incomplete: ... - def getiterator(self, tag: Incomplete | None = None) -> Incomplete: ... - def makeelement( - self, tag: Incomplete, attrib: Incomplete - ) -> Incomplete: ... - def find(self, path: Incomplete) -> Incomplete: ... - def findall(self, path: Incomplete) -> Incomplete: ... - -def raise_unorderable_types( - ordering: Incomplete, a: Incomplete, b: Incomplete -) -> None: ... diff --git a/stubs/nltk/jsontags.pyi b/stubs/nltk/jsontags.pyi deleted file mode 100644 index de48870..0000000 --- a/stubs/nltk/jsontags.pyi +++ /dev/null @@ -1,22 +0,0 @@ -import json - -from _typeshed import Incomplete - -__all__ = [ - "register_tag", - "json_tags", - "JSONTaggedEncoder", - "JSONTaggedDecoder", -] - -json_tags: Incomplete - -def register_tag(cls: Incomplete) -> Incomplete: ... - -class JSONTaggedEncoder(json.JSONEncoder): - def default(self, obj: Incomplete) -> Incomplete: ... - -class JSONTaggedDecoder(json.JSONDecoder): - def decode(self, s: Incomplete) -> Incomplete: ... - @classmethod - def decode_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/langnames.pyi b/stubs/nltk/langnames.pyi deleted file mode 100644 index 2c0797c..0000000 --- a/stubs/nltk/langnames.pyi +++ /dev/null @@ -1,19 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus import bcp47 as bcp47 - -codepattern: Incomplete - -def langname(tag: Incomplete, typ: str = "full") -> Incomplete: ... -def langcode(name: Incomplete, typ: int = 2) -> Incomplete: ... -def tag2q(tag: Incomplete) -> Incomplete: ... -def q2tag(qcode: Incomplete) -> Incomplete: ... -def q2name(qcode: Incomplete, typ: str = "full") -> Incomplete: ... -def lang2q(name: Incomplete) -> Incomplete: ... -def inverse_dict(dic: Incomplete) -> Incomplete: ... - -wiki_bcp47: Incomplete -iso639short: Incomplete -iso639retired: Incomplete -iso639long: Incomplete -iso639code_retired: Incomplete diff --git a/stubs/nltk/lazyimport.pyi b/stubs/nltk/lazyimport.pyi deleted file mode 100644 index a3b2b6c..0000000 --- a/stubs/nltk/lazyimport.pyi +++ /dev/null @@ -1,11 +0,0 @@ -from _typeshed import Incomplete - -class LazyModule: - def __init__( - self, - name: Incomplete, - locals: Incomplete, - globals: Incomplete | None = None, - ) -> None: ... - def __getattr__(self, name: Incomplete) -> Incomplete: ... - def __setattr__(self, name: Incomplete, value: Incomplete) -> None: ... diff --git a/stubs/nltk/lm/__init__.pyi b/stubs/nltk/lm/__init__.pyi deleted file mode 100644 index cb41855..0000000 --- a/stubs/nltk/lm/__init__.pyi +++ /dev/null @@ -1,35 +0,0 @@ -from nltk.lm.counter import NgramCounter as NgramCounter -from nltk.lm.models import ( - MLE as MLE, -) -from nltk.lm.models import ( - AbsoluteDiscountingInterpolated as AbsoluteDiscountingInterpolated, -) -from nltk.lm.models import ( - KneserNeyInterpolated as KneserNeyInterpolated, -) -from nltk.lm.models import ( - Laplace as Laplace, -) -from nltk.lm.models import ( - Lidstone as Lidstone, -) -from nltk.lm.models import ( - StupidBackoff as StupidBackoff, -) -from nltk.lm.models import ( - WittenBellInterpolated as WittenBellInterpolated, -) -from nltk.lm.vocabulary import Vocabulary as Vocabulary - -__all__ = [ - "Vocabulary", - "NgramCounter", - "MLE", - "Lidstone", - "Laplace", - "WittenBellInterpolated", - "KneserNeyInterpolated", - "AbsoluteDiscountingInterpolated", - "StupidBackoff", -] diff --git a/stubs/nltk/lm/api.pyi b/stubs/nltk/lm/api.pyi deleted file mode 100644 index eb75389..0000000 --- a/stubs/nltk/lm/api.pyi +++ /dev/null @@ -1,53 +0,0 @@ -from abc import ABCMeta, abstractmethod - -from _typeshed import Incomplete - -from nltk.lm.counter import NgramCounter as NgramCounter -from nltk.lm.util import log_base2 as log_base2 -from nltk.lm.vocabulary import Vocabulary as Vocabulary - -class Smoothing(metaclass=ABCMeta): - vocab: Incomplete - counts: Incomplete - def __init__( - self, vocabulary: Incomplete, counter: Incomplete - ) -> None: ... - @abstractmethod - def unigram_score(self, word: Incomplete) -> Incomplete: ... - @abstractmethod - def alpha_gamma( - self, word: Incomplete, context: Incomplete - ) -> Incomplete: ... - -class LanguageModel(metaclass=ABCMeta): - order: Incomplete - vocab: Incomplete - counts: Incomplete - def __init__( - self, - order: Incomplete, - vocabulary: Incomplete | None = None, - counter: Incomplete | None = None, - ) -> None: ... - def fit( - self, text: Incomplete, vocabulary_text: Incomplete | None = None - ) -> None: ... - def score( - self, word: Incomplete, context: Incomplete | None = None - ) -> Incomplete: ... - @abstractmethod - def unmasked_score( - self, word: Incomplete, context: Incomplete | None = None - ) -> Incomplete: ... - def logscore( - self, word: Incomplete, context: Incomplete | None = None - ) -> Incomplete: ... - def context_counts(self, context: Incomplete) -> Incomplete: ... - def entropy(self, text_ngrams: Incomplete) -> Incomplete: ... - def perplexity(self, text_ngrams: Incomplete) -> Incomplete: ... - def generate( - self, - num_words: int = 1, - text_seed: Incomplete | None = None, - random_seed: Incomplete | None = None, - ) -> Incomplete: ... diff --git a/stubs/nltk/lm/counter.pyi b/stubs/nltk/lm/counter.pyi deleted file mode 100644 index 32c91ab..0000000 --- a/stubs/nltk/lm/counter.pyi +++ /dev/null @@ -1,16 +0,0 @@ -from _typeshed import Incomplete - -from nltk.probability import ( - ConditionalFreqDist as ConditionalFreqDist, -) -from nltk.probability import ( - FreqDist as FreqDist, -) - -class NgramCounter: - def __init__(self, ngram_text: Incomplete | None = None) -> None: ... - def update(self, ngram_text: Incomplete) -> None: ... - def N(self) -> Incomplete: ... - def __getitem__(self, item: Incomplete) -> Incomplete: ... - def __len__(self) -> int: ... - def __contains__(self, item: Incomplete) -> bool: ... diff --git a/stubs/nltk/lm/models.pyi b/stubs/nltk/lm/models.pyi deleted file mode 100644 index f249fd1..0000000 --- a/stubs/nltk/lm/models.pyi +++ /dev/null @@ -1,64 +0,0 @@ -from _typeshed import Incomplete - -from nltk.lm.api import LanguageModel as LanguageModel -from nltk.lm.api import Smoothing as Smoothing -from nltk.lm.smoothing import ( - AbsoluteDiscounting as AbsoluteDiscounting, -) -from nltk.lm.smoothing import ( - KneserNey as KneserNey, -) -from nltk.lm.smoothing import ( - WittenBell as WittenBell, -) - -class MLE(LanguageModel): - def unmasked_score( - self, word: Incomplete, context: Incomplete | None = None - ) -> Incomplete: ... - -class Lidstone(LanguageModel): - gamma: Incomplete - def __init__( - self, gamma: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - def unmasked_score( - self, word: Incomplete, context: Incomplete | None = None - ) -> Incomplete: ... - -class Laplace(Lidstone): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class StupidBackoff(LanguageModel): - alpha: Incomplete - def __init__( - self, alpha: float = 0.4, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - def unmasked_score( - self, word: Incomplete, context: Incomplete | None = None - ) -> Incomplete: ... - -class InterpolatedLanguageModel(LanguageModel): - estimator: Incomplete - def __init__( - self, - smoothing_cls: Incomplete, - order: Incomplete, - **kwargs: Incomplete, - ) -> None: ... - def unmasked_score( - self, word: Incomplete, context: Incomplete | None = None - ) -> Incomplete: ... - -class WittenBellInterpolated(InterpolatedLanguageModel): - def __init__(self, order: Incomplete, **kwargs: Incomplete) -> None: ... - -class AbsoluteDiscountingInterpolated(InterpolatedLanguageModel): - def __init__( - self, order: Incomplete, discount: float = 0.75, **kwargs: Incomplete - ) -> None: ... - -class KneserNeyInterpolated(InterpolatedLanguageModel): - def __init__( - self, order: Incomplete, discount: float = 0.1, **kwargs: Incomplete - ) -> None: ... diff --git a/stubs/nltk/lm/preprocessing.pyi b/stubs/nltk/lm/preprocessing.pyi deleted file mode 100644 index cc181d9..0000000 --- a/stubs/nltk/lm/preprocessing.pyi +++ /dev/null @@ -1,14 +0,0 @@ -from _typeshed import Incomplete - -from nltk.util import everygrams as everygrams -from nltk.util import pad_sequence as pad_sequence - -flatten: Incomplete -pad_both_ends: Incomplete - -def padded_everygrams( - order: Incomplete, sentence: Incomplete -) -> Incomplete: ... -def padded_everygram_pipeline( - order: Incomplete, text: Incomplete -) -> Incomplete: ... diff --git a/stubs/nltk/lm/smoothing.pyi b/stubs/nltk/lm/smoothing.pyi deleted file mode 100644 index 9a07679..0000000 --- a/stubs/nltk/lm/smoothing.pyi +++ /dev/null @@ -1,42 +0,0 @@ -from _typeshed import Incomplete - -from nltk.lm.api import Smoothing as Smoothing -from nltk.probability import ConditionalFreqDist as ConditionalFreqDist - -class WittenBell(Smoothing): - def __init__( - self, vocabulary: Incomplete, counter: Incomplete, **kwargs: Incomplete - ) -> None: ... - def alpha_gamma( - self, word: Incomplete, context: Incomplete - ) -> Incomplete: ... - def unigram_score(self, word: Incomplete) -> Incomplete: ... - -class AbsoluteDiscounting(Smoothing): - discount: Incomplete - def __init__( - self, - vocabulary: Incomplete, - counter: Incomplete, - discount: float = 0.75, - **kwargs: Incomplete, - ) -> None: ... - def alpha_gamma( - self, word: Incomplete, context: Incomplete - ) -> Incomplete: ... - def unigram_score(self, word: Incomplete) -> Incomplete: ... - -class KneserNey(Smoothing): - discount: Incomplete - def __init__( - self, - vocabulary: Incomplete, - counter: Incomplete, - order: Incomplete, - discount: float = 0.1, - **kwargs: Incomplete, - ) -> None: ... - def unigram_score(self, word: Incomplete) -> Incomplete: ... - def alpha_gamma( - self, word: Incomplete, context: Incomplete - ) -> Incomplete: ... diff --git a/stubs/nltk/lm/util.pyi b/stubs/nltk/lm/util.pyi deleted file mode 100644 index 1d2c14a..0000000 --- a/stubs/nltk/lm/util.pyi +++ /dev/null @@ -1,6 +0,0 @@ -from _typeshed import Incomplete - -NEG_INF: Incomplete -POS_INF: Incomplete - -def log_base2(score: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/lm/vocabulary.pyi b/stubs/nltk/lm/vocabulary.pyi deleted file mode 100644 index 240df80..0000000 --- a/stubs/nltk/lm/vocabulary.pyi +++ /dev/null @@ -1,24 +0,0 @@ -from _typeshed import Incomplete - -def _(words: Incomplete, vocab: Incomplete) -> Incomplete: ... - -class Vocabulary: - unk_label: Incomplete - counts: Incomplete - def __init__( - self, - counts: Incomplete | None = None, - unk_cutoff: int = 1, - unk_label: str = "", - ) -> None: ... - @property - def cutoff(self) -> Incomplete: ... - def update( - self, *counter_args: Incomplete, **counter_kwargs: Incomplete - ) -> None: ... - def lookup(self, words: Incomplete) -> Incomplete: ... - def __getitem__(self, item: Incomplete) -> Incomplete: ... - def __contains__(self, item: Incomplete) -> bool: ... - def __iter__(self) -> Incomplete: ... - def __len__(self) -> int: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/metrics/__init__.pyi b/stubs/nltk/metrics/__init__.pyi deleted file mode 100644 index 32b799e..0000000 --- a/stubs/nltk/metrics/__init__.pyi +++ /dev/null @@ -1,82 +0,0 @@ -from nltk.metrics.agreement import AnnotationTask as AnnotationTask -from nltk.metrics.aline import align as align -from nltk.metrics.association import ( - BigramAssocMeasures as BigramAssocMeasures, -) -from nltk.metrics.association import ( - ContingencyMeasures as ContingencyMeasures, -) -from nltk.metrics.association import ( - NgramAssocMeasures as NgramAssocMeasures, -) -from nltk.metrics.association import ( - QuadgramAssocMeasures as QuadgramAssocMeasures, -) -from nltk.metrics.association import ( - TrigramAssocMeasures as TrigramAssocMeasures, -) -from nltk.metrics.confusionmatrix import ConfusionMatrix as ConfusionMatrix -from nltk.metrics.distance import ( - binary_distance as binary_distance, -) -from nltk.metrics.distance import ( - custom_distance as custom_distance, -) -from nltk.metrics.distance import ( - edit_distance as edit_distance, -) -from nltk.metrics.distance import ( - edit_distance_align as edit_distance_align, -) -from nltk.metrics.distance import ( - fractional_presence as fractional_presence, -) -from nltk.metrics.distance import ( - interval_distance as interval_distance, -) -from nltk.metrics.distance import ( - jaccard_distance as jaccard_distance, -) -from nltk.metrics.distance import ( - masi_distance as masi_distance, -) -from nltk.metrics.distance import ( - presence as presence, -) -from nltk.metrics.paice import Paice as Paice -from nltk.metrics.scores import ( - accuracy as accuracy, -) -from nltk.metrics.scores import ( - approxrand as approxrand, -) -from nltk.metrics.scores import ( - f_measure as f_measure, -) -from nltk.metrics.scores import ( - log_likelihood as log_likelihood, -) -from nltk.metrics.scores import ( - precision as precision, -) -from nltk.metrics.scores import ( - recall as recall, -) -from nltk.metrics.segmentation import ( - ghd as ghd, -) -from nltk.metrics.segmentation import ( - pk as pk, -) -from nltk.metrics.segmentation import ( - windowdiff as windowdiff, -) -from nltk.metrics.spearman import ( - ranks_from_scores as ranks_from_scores, -) -from nltk.metrics.spearman import ( - ranks_from_sequence as ranks_from_sequence, -) -from nltk.metrics.spearman import ( - spearman_correlation as spearman_correlation, -) diff --git a/stubs/nltk/metrics/agreement.pyi b/stubs/nltk/metrics/agreement.pyi deleted file mode 100644 index 08e63a9..0000000 --- a/stubs/nltk/metrics/agreement.pyi +++ /dev/null @@ -1,57 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import deprecated as deprecated -from nltk.metrics.distance import binary_distance as binary_distance -from nltk.probability import ( - ConditionalFreqDist as ConditionalFreqDist, -) -from nltk.probability import ( - FreqDist as FreqDist, -) - -log: Incomplete - -class AnnotationTask: - distance: Incomplete - I: Incomplete - K: Incomplete - C: Incomplete - data: Incomplete - def __init__( - self, data: Incomplete | None = None, distance: Incomplete = ... - ) -> None: ... - def load_array(self, array: Incomplete) -> None: ... - def agr( - self, - cA: Incomplete, - cB: Incomplete, - i: Incomplete, - data: Incomplete | None = None, - ) -> Incomplete: ... - def Nk(self, k: Incomplete) -> Incomplete: ... - def Nik(self, i: Incomplete, k: Incomplete) -> Incomplete: ... - def Nck(self, c: Incomplete, k: Incomplete) -> Incomplete: ... - def N( - self, - k: Incomplete | None = None, - i: Incomplete | None = None, - c: Incomplete | None = None, - ) -> Incomplete: ... - def Ao(self, cA: Incomplete, cB: Incomplete) -> Incomplete: ... - def avg_Ao(self) -> Incomplete: ... - def Do_Kw_pairwise( - self, cA: Incomplete, cB: Incomplete, max_distance: float = 1.0 - ) -> Incomplete: ... - def Do_Kw(self, max_distance: float = 1.0) -> Incomplete: ... - def S(self) -> Incomplete: ... - def pi(self) -> Incomplete: ... - def Ae_kappa(self, cA: Incomplete, cB: Incomplete) -> Incomplete: ... - def kappa_pairwise(self, cA: Incomplete, cB: Incomplete) -> Incomplete: ... - def kappa(self) -> Incomplete: ... - def multi_kappa(self) -> Incomplete: ... - def Disagreement(self, label_freqs: Incomplete) -> Incomplete: ... - def alpha(self) -> Incomplete: ... - def weighted_kappa_pairwise( - self, cA: Incomplete, cB: Incomplete, max_distance: float = 1.0 - ) -> Incomplete: ... - def weighted_kappa(self, max_distance: float = 1.0) -> Incomplete: ... diff --git a/stubs/nltk/metrics/aline.pyi b/stubs/nltk/metrics/aline.pyi deleted file mode 100644 index 070da03..0000000 --- a/stubs/nltk/metrics/aline.pyi +++ /dev/null @@ -1,47 +0,0 @@ -from typing import ( - Any, - List, - Tuple, - Union, -) - -from _typeshed import Incomplete -from numpy import ( - float64, - ndarray, -) - -inf: Incomplete -C_skip: int -C_sub: int -C_exp: int -C_vwl: int -consonants: Incomplete -R_c: Incomplete -R_v: Incomplete -similarity_matrix: Incomplete -salience: Incomplete -feature_matrix: Incomplete - -def R(p: str, q: str) -> List[str]: ... -def V(p: str) -> int: ... -def _retrieve( - i: int, - j: int, - s: Union[int, float64], - S: ndarray, - T: float64, - str1: str, - str2: str, - out: List[Union[Any, Tuple[str, str]]], -) -> List[Tuple[str, str]]: ... -def align( - str1: str, str2: str, epsilon: int = ... -) -> List[List[Tuple[str, str]]]: ... -def delta(p: str, q: str) -> float64: ... -def diff(p: str, q: str, f: str) -> float: ... -def sigma_exp(p: str, q: str) -> float64: ... -def sigma_skip(p: str) -> int: ... -def sigma_sub(p: str, q: str) -> float64: ... - -cognate_data: str diff --git a/stubs/nltk/metrics/association.pyi b/stubs/nltk/metrics/association.pyi deleted file mode 100644 index 7f5a7b0..0000000 --- a/stubs/nltk/metrics/association.pyi +++ /dev/null @@ -1,89 +0,0 @@ -from typing import Tuple - -from _typeshed import Incomplete - -NGRAM: int -UNIGRAMS: int -TOTAL: int - -class BigramAssocMeasures: - @staticmethod - def _contingency( - n_ii: int, n_ix_xi_tuple: Tuple[int, int], n_xx: int - ) -> Tuple[int, int, int, int]: ... - @classmethod - def phi_sq(cls: Incomplete, *marginals: Incomplete) -> Incomplete: ... - @classmethod - def chi_sq( - cls: Incomplete, - n_ii: Incomplete, - n_ix_xi_tuple: Incomplete, - n_xx: Incomplete, - ) -> Incomplete: ... - @classmethod - def fisher(cls: Incomplete, *marginals: Incomplete) -> Incomplete: ... - @staticmethod - def dice( - n_ii: Incomplete, n_ix_xi_tuple: Incomplete, n_xx: Incomplete - ) -> Incomplete: ... - -class NgramAssocMeasures: - @classmethod - def likelihood_ratio(cls: Incomplete, *marginals: Incomplete) -> float: ... - @classmethod - def pmi(cls: Incomplete, *marginals: Incomplete) -> float: ... - @staticmethod - def raw_freq(*marginals: Incomplete) -> Incomplete: ... - @classmethod - def student_t(cls: Incomplete, *marginals: Incomplete) -> Incomplete: ... - @classmethod - def chi_sq(cls: Incomplete, *marginals: Incomplete) -> Incomplete: ... - @staticmethod - def mi_like( - *marginals: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - @classmethod - def poisson_stirling( - cls: Incomplete, *marginals: Incomplete - ) -> Incomplete: ... - @classmethod - def jaccard(cls: Incomplete, *marginals: Incomplete) -> Incomplete: ... - -class QuadgramAssocMeasures: - @staticmethod - def _contingency( - n_iiii: int, - n_iiix_tuple: Tuple[int, int, int, int], - n_iixx_tuple: Tuple[int, int, int, int, int, int], - n_ixxx_tuple: Tuple[int, int, int, int], - n_xxxx: int, - ) -> Tuple[ - int, - int, - int, - int, - int, - int, - int, - int, - int, - int, - int, - int, - int, - int, - int, - int, - ]: ... - -class TrigramAssocMeasures: - @staticmethod - def _contingency( - n_iii: int, - n_iix_tuple: Tuple[int, int, int], - n_ixx_tuple: Tuple[int, int, int], - n_xxx: int, - ) -> Tuple[int, int, int, int, int, int, int, int]: ... - -class ContingencyMeasures: - def __init__(self, measures: Incomplete) -> None: ... diff --git a/stubs/nltk/metrics/confusionmatrix.pyi b/stubs/nltk/metrics/confusionmatrix.pyi deleted file mode 100644 index c0dedf0..0000000 --- a/stubs/nltk/metrics/confusionmatrix.pyi +++ /dev/null @@ -1,33 +0,0 @@ -from _typeshed import Incomplete - -from nltk.probability import FreqDist as FreqDist - -class ConfusionMatrix: - def __init__( - self, - reference: Incomplete, - test: Incomplete, - sort_by_count: bool = False, - ) -> None: ... - def __getitem__(self, li_lj_tuple: Incomplete) -> Incomplete: ... - def pretty_format( - self, - show_percents: bool = False, - values_in_chart: bool = True, - truncate: Incomplete | None = None, - sort_by_count: bool = False, - ) -> Incomplete: ... - def key(self) -> Incomplete: ... - def recall(self, value: Incomplete) -> Incomplete: ... - def precision(self, value: Incomplete) -> Incomplete: ... - def f_measure( - self, value: Incomplete, alpha: float = 0.5 - ) -> Incomplete: ... - def evaluate( - self, - alpha: float = 0.5, - truncate: Incomplete | None = None, - sort_by_count: bool = False, - ) -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/metrics/distance.pyi b/stubs/nltk/metrics/distance.pyi deleted file mode 100644 index dccd030..0000000 --- a/stubs/nltk/metrics/distance.pyi +++ /dev/null @@ -1,29 +0,0 @@ -from typing import Union - -from _typeshed import Incomplete - -def binary_distance( - label1: Union[int, str], label2: Union[int, str] -) -> float: ... -def edit_distance( - s1: Incomplete, - s2: Incomplete, - substitution_cost: int = 1, - transpositions: bool = False, -) -> Incomplete: ... -def edit_distance_align( - s1: Incomplete, s2: Incomplete, substitution_cost: int = 1 -) -> Incomplete: ... -def jaccard_distance(label1: Incomplete, label2: Incomplete) -> Incomplete: ... -def masi_distance(label1: Incomplete, label2: Incomplete) -> Incomplete: ... -def interval_distance( - label1: Incomplete, label2: Incomplete -) -> Incomplete: ... -def presence(label: Incomplete) -> Incomplete: ... -def fractional_presence(label: Incomplete) -> Incomplete: ... -def custom_distance(file: Incomplete) -> Incomplete: ... -def jaro_similarity(s1: Incomplete, s2: Incomplete) -> Incomplete: ... -def jaro_winkler_similarity( - s1: Incomplete, s2: Incomplete, p: float = 0.1, max_l: int = 4 -) -> Incomplete: ... -def demo() -> None: ... diff --git a/stubs/nltk/metrics/paice.pyi b/stubs/nltk/metrics/paice.pyi deleted file mode 100644 index e616252..0000000 --- a/stubs/nltk/metrics/paice.pyi +++ /dev/null @@ -1,13 +0,0 @@ -from _typeshed import Incomplete - -def get_words_from_dictionary(lemmas: Incomplete) -> Incomplete: ... - -class Paice: - lemmas: Incomplete - stems: Incomplete - coords: Incomplete - errt: Incomplete - def __init__(self, lemmas: Incomplete, stems: Incomplete) -> None: ... - def update(self) -> None: ... - -def demo() -> None: ... diff --git a/stubs/nltk/metrics/scores.pyi b/stubs/nltk/metrics/scores.pyi deleted file mode 100644 index 35384fc..0000000 --- a/stubs/nltk/metrics/scores.pyi +++ /dev/null @@ -1,20 +0,0 @@ -from _typeshed import Incomplete - -from nltk.util import ( - LazyConcatenation as LazyConcatenation, -) -from nltk.util import ( - LazyMap as LazyMap, -) - -def accuracy(reference: Incomplete, test: Incomplete) -> Incomplete: ... -def precision(reference: Incomplete, test: Incomplete) -> Incomplete: ... -def recall(reference: Incomplete, test: Incomplete) -> Incomplete: ... -def f_measure( - reference: Incomplete, test: Incomplete, alpha: float = 0.5 -) -> Incomplete: ... -def log_likelihood(reference: Incomplete, test: Incomplete) -> Incomplete: ... -def approxrand( - a: Incomplete, b: Incomplete, **kwargs: Incomplete -) -> Incomplete: ... -def demo() -> None: ... diff --git a/stubs/nltk/metrics/segmentation.pyi b/stubs/nltk/metrics/segmentation.pyi deleted file mode 100644 index c977155..0000000 --- a/stubs/nltk/metrics/segmentation.pyi +++ /dev/null @@ -1,23 +0,0 @@ -from _typeshed import Incomplete - -def windowdiff( - seg1: Incomplete, - seg2: Incomplete, - k: Incomplete, - boundary: str = "1", - weighted: bool = False, -) -> Incomplete: ... -def ghd( - ref: Incomplete, - hyp: Incomplete, - ins_cost: float = 2.0, - del_cost: float = 2.0, - shift_cost_coeff: float = 1.0, - boundary: str = "1", -) -> Incomplete: ... -def pk( - ref: Incomplete, - hyp: Incomplete, - k: Incomplete | None = None, - boundary: str = "1", -) -> Incomplete: ... diff --git a/stubs/nltk/metrics/spearman.pyi b/stubs/nltk/metrics/spearman.pyi deleted file mode 100644 index 4298827..0000000 --- a/stubs/nltk/metrics/spearman.pyi +++ /dev/null @@ -1,11 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -def spearman_correlation( - ranks1: Incomplete, ranks2: Incomplete -) -> Incomplete: ... -def ranks_from_sequence(seq: Incomplete) -> Incomplete: ... -def ranks_from_scores( - scores: Incomplete, rank_gap: float = 1e-15 -) -> Generator[Incomplete, None, None]: ... diff --git a/stubs/nltk/misc/__init__.pyi b/stubs/nltk/misc/__init__.pyi deleted file mode 100644 index d8a073b..0000000 --- a/stubs/nltk/misc/__init__.pyi +++ /dev/null @@ -1,4 +0,0 @@ -from nltk.misc.babelfish import babelize_shell as babelize_shell -from nltk.misc.chomsky import generate_chomsky as generate_chomsky -from nltk.misc.minimalset import MinimalSet as MinimalSet -from nltk.misc.wordfinder import word_finder as word_finder diff --git a/stubs/nltk/misc/babelfish.pyi b/stubs/nltk/misc/babelfish.pyi deleted file mode 100644 index 763e7aa..0000000 --- a/stubs/nltk/misc/babelfish.pyi +++ /dev/null @@ -1 +0,0 @@ -def babelize_shell() -> None: ... diff --git a/stubs/nltk/misc/chomsky.pyi b/stubs/nltk/misc/chomsky.pyi deleted file mode 100644 index 126a89c..0000000 --- a/stubs/nltk/misc/chomsky.pyi +++ /dev/null @@ -1,6 +0,0 @@ -leadins: str -subjects: str -verbs: str -objects: str - -def generate_chomsky(times: int = 5, line_length: int = 72) -> None: ... diff --git a/stubs/nltk/misc/minimalset.pyi b/stubs/nltk/misc/minimalset.pyi deleted file mode 100644 index f0c5138..0000000 --- a/stubs/nltk/misc/minimalset.pyi +++ /dev/null @@ -1,13 +0,0 @@ -from _typeshed import Incomplete - -class MinimalSet: - def __init__(self, parameters: Incomplete | None = None) -> None: ... - def add( - self, context: Incomplete, target: Incomplete, display: Incomplete - ) -> None: ... - def contexts(self, minimum: int = 2) -> Incomplete: ... - def display( - self, context: Incomplete, target: Incomplete, default: str = "" - ) -> Incomplete: ... - def display_all(self, context: Incomplete) -> Incomplete: ... - def targets(self) -> Incomplete: ... diff --git a/stubs/nltk/misc/sort.pyi b/stubs/nltk/misc/sort.pyi deleted file mode 100644 index 1f9c5d5..0000000 --- a/stubs/nltk/misc/sort.pyi +++ /dev/null @@ -1,7 +0,0 @@ -from _typeshed import Incomplete - -def selection(a: Incomplete) -> Incomplete: ... -def bubble(a: Incomplete) -> Incomplete: ... -def merge(a: Incomplete) -> Incomplete: ... -def quick(a: Incomplete) -> Incomplete: ... -def demo() -> None: ... diff --git a/stubs/nltk/misc/wordfinder.pyi b/stubs/nltk/misc/wordfinder.pyi deleted file mode 100644 index bd520e6..0000000 --- a/stubs/nltk/misc/wordfinder.pyi +++ /dev/null @@ -1,28 +0,0 @@ -from _typeshed import Incomplete - -def revword(word: Incomplete) -> Incomplete: ... -def step( - word: Incomplete, - x: Incomplete, - xf: Incomplete, - y: Incomplete, - yf: Incomplete, - grid: Incomplete, -) -> Incomplete: ... -def check( - word: Incomplete, - dir: Incomplete, - x: Incomplete, - y: Incomplete, - grid: Incomplete, - rows: Incomplete, - cols: Incomplete, -) -> Incomplete: ... -def wordfinder( - words: Incomplete, - rows: int = 20, - cols: int = 20, - attempts: int = 50, - alph: str = "ABCDEFGHIJKLMNOPQRSTUVWXYZ", -) -> Incomplete: ... -def word_finder() -> None: ... diff --git a/stubs/nltk/parse/__init__.pyi b/stubs/nltk/parse/__init__.pyi deleted file mode 100644 index 93526df..0000000 --- a/stubs/nltk/parse/__init__.pyi +++ /dev/null @@ -1,127 +0,0 @@ -from nltk.parse.api import ParserI as ParserI -from nltk.parse.bllip import BllipParser as BllipParser -from nltk.parse.chart import ( - BottomUpChartParser as BottomUpChartParser, -) -from nltk.parse.chart import ( - BottomUpLeftCornerChartParser as BottomUpLeftCornerChartParser, -) -from nltk.parse.chart import ( - ChartParser as ChartParser, -) -from nltk.parse.chart import ( - LeftCornerChartParser as LeftCornerChartParser, -) -from nltk.parse.chart import ( - SteppingChartParser as SteppingChartParser, -) -from nltk.parse.chart import ( - TopDownChartParser as TopDownChartParser, -) -from nltk.parse.corenlp import ( - CoreNLPDependencyParser as CoreNLPDependencyParser, -) -from nltk.parse.corenlp import ( - CoreNLPParser as CoreNLPParser, -) -from nltk.parse.dependencygraph import DependencyGraph as DependencyGraph -from nltk.parse.earleychart import ( - EarleyChartParser as EarleyChartParser, -) -from nltk.parse.earleychart import ( - FeatureEarleyChartParser as FeatureEarleyChartParser, -) -from nltk.parse.earleychart import ( - FeatureIncrementalBottomUpChartParser as FeatureIncrementalBottomUpChartParser, -) -from nltk.parse.earleychart import ( - FeatureIncrementalBottomUpLeftCornerChartParser as FeatureIncrementalBottomUpLeftCornerChartParser, -) -from nltk.parse.earleychart import ( - FeatureIncrementalChartParser as FeatureIncrementalChartParser, -) -from nltk.parse.earleychart import ( - FeatureIncrementalTopDownChartParser as FeatureIncrementalTopDownChartParser, -) -from nltk.parse.earleychart import ( - IncrementalBottomUpChartParser as IncrementalBottomUpChartParser, -) -from nltk.parse.earleychart import ( - IncrementalBottomUpLeftCornerChartParser as IncrementalBottomUpLeftCornerChartParser, -) -from nltk.parse.earleychart import ( - IncrementalChartParser as IncrementalChartParser, -) -from nltk.parse.earleychart import ( - IncrementalLeftCornerChartParser as IncrementalLeftCornerChartParser, -) -from nltk.parse.earleychart import ( - IncrementalTopDownChartParser as IncrementalTopDownChartParser, -) -from nltk.parse.evaluate import DependencyEvaluator as DependencyEvaluator -from nltk.parse.featurechart import ( - FeatureBottomUpChartParser as FeatureBottomUpChartParser, -) -from nltk.parse.featurechart import ( - FeatureBottomUpLeftCornerChartParser as FeatureBottomUpLeftCornerChartParser, -) -from nltk.parse.featurechart import ( - FeatureChartParser as FeatureChartParser, -) -from nltk.parse.featurechart import ( - FeatureTopDownChartParser as FeatureTopDownChartParser, -) -from nltk.parse.malt import MaltParser as MaltParser -from nltk.parse.nonprojectivedependencyparser import ( - NaiveBayesDependencyScorer as NaiveBayesDependencyScorer, -) -from nltk.parse.nonprojectivedependencyparser import ( - NonprojectiveDependencyParser as NonprojectiveDependencyParser, -) -from nltk.parse.nonprojectivedependencyparser import ( - ProbabilisticNonprojectiveParser as ProbabilisticNonprojectiveParser, -) -from nltk.parse.pchart import ( - BottomUpProbabilisticChartParser as BottomUpProbabilisticChartParser, -) -from nltk.parse.pchart import ( - InsideChartParser as InsideChartParser, -) -from nltk.parse.pchart import ( - LongestChartParser as LongestChartParser, -) -from nltk.parse.pchart import ( - RandomChartParser as RandomChartParser, -) -from nltk.parse.pchart import ( - UnsortedChartParser as UnsortedChartParser, -) -from nltk.parse.projectivedependencyparser import ( - ProbabilisticProjectiveDependencyParser as ProbabilisticProjectiveDependencyParser, -) -from nltk.parse.projectivedependencyparser import ( - ProjectiveDependencyParser as ProjectiveDependencyParser, -) -from nltk.parse.recursivedescent import ( - RecursiveDescentParser as RecursiveDescentParser, -) -from nltk.parse.recursivedescent import ( - SteppingRecursiveDescentParser as SteppingRecursiveDescentParser, -) -from nltk.parse.shiftreduce import ( - ShiftReduceParser as ShiftReduceParser, -) -from nltk.parse.shiftreduce import ( - SteppingShiftReduceParser as SteppingShiftReduceParser, -) -from nltk.parse.transitionparser import TransitionParser as TransitionParser -from nltk.parse.util import ( - TestGrammar as TestGrammar, -) -from nltk.parse.util import ( - extract_test_sentences as extract_test_sentences, -) -from nltk.parse.util import ( - load_parser as load_parser, -) -from nltk.parse.viterbi import ViterbiParser as ViterbiParser diff --git a/stubs/nltk/parse/api.pyi b/stubs/nltk/parse/api.pyi deleted file mode 100644 index d7e296b..0000000 --- a/stubs/nltk/parse/api.pyi +++ /dev/null @@ -1,18 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import overridden as overridden - -class ParserI: - def grammar(self) -> None: ... - def parse( - self, sent: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def parse_sents( - self, sents: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def parse_all( - self, sent: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def parse_one( - self, sent: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... diff --git a/stubs/nltk/parse/bllip.pyi b/stubs/nltk/parse/bllip.pyi deleted file mode 100644 index de1b738..0000000 --- a/stubs/nltk/parse/bllip.pyi +++ /dev/null @@ -1,31 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.parse.api import ParserI - -__all__ = ["BllipParser"] - -class BllipParser(ParserI): - rrp: Incomplete - def __init__( - self, - parser_model: Incomplete | None = None, - reranker_features: Incomplete | None = None, - reranker_weights: Incomplete | None = None, - parser_options: Incomplete | None = None, - reranker_options: Incomplete | None = None, - ) -> None: ... - def parse( - self, sentence: Incomplete - ) -> Generator[Incomplete, None, None]: ... - def tagged_parse( - self, word_and_tag_pairs: Incomplete - ) -> Generator[Incomplete, None, None]: ... - @classmethod - def from_unified_model_dir( - cls: Incomplete, - model_dir: Incomplete, - parser_options: Incomplete | None = None, - reranker_options: Incomplete | None = None, - ) -> Incomplete: ... diff --git a/stubs/nltk/parse/chart.pyi b/stubs/nltk/parse/chart.pyi deleted file mode 100644 index 316dc48..0000000 --- a/stubs/nltk/parse/chart.pyi +++ /dev/null @@ -1,261 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.grammar import PCFG as PCFG -from nltk.grammar import is_nonterminal as is_nonterminal -from nltk.grammar import is_terminal as is_terminal -from nltk.internals import raise_unorderable_types as raise_unorderable_types -from nltk.parse.api import ParserI as ParserI -from nltk.tree import Tree as Tree -from nltk.util import OrderedDict as OrderedDict - -class EdgeI: - def __init__(self) -> None: ... - def span(self) -> None: ... - def start(self) -> None: ... - def end(self) -> None: ... - def length(self) -> None: ... - def lhs(self) -> None: ... - def rhs(self) -> None: ... - def dot(self) -> None: ... - def nextsym(self) -> None: ... - def is_complete(self) -> None: ... - def is_incomplete(self) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class TreeEdge(EdgeI): - def __init__( - self, span: Incomplete, lhs: Incomplete, rhs: Incomplete, dot: int = 0 - ) -> None: ... - @staticmethod - def from_production( - production: Incomplete, index: Incomplete - ) -> Incomplete: ... - def move_dot_forward(self, new_en: Incomplete) -> Incomplete: ... - def lhs(self) -> Incomplete: ... - def span(self) -> Incomplete: ... - def start(self) -> Incomplete: ... - def end(self) -> Incomplete: ... - def length(self) -> Incomplete: ... - def rhs(self) -> Incomplete: ... - def dot(self) -> Incomplete: ... - def is_complete(self) -> Incomplete: ... - def is_incomplete(self) -> Incomplete: ... - def nextsym(self) -> Incomplete: ... - -class LeafEdge(EdgeI): - def __init__(self, leaf: Incomplete, index: Incomplete) -> None: ... - def lhs(self) -> Incomplete: ... - def span(self) -> Incomplete: ... - def start(self) -> Incomplete: ... - def end(self) -> Incomplete: ... - def length(self) -> Incomplete: ... - def rhs(self) -> Incomplete: ... - def dot(self) -> Incomplete: ... - def is_complete(self) -> Incomplete: ... - def is_incomplete(self) -> Incomplete: ... - def nextsym(self) -> None: ... - -class Chart: - def __init__(self, tokens: Incomplete) -> None: ... - def initialize(self) -> None: ... - def num_leaves(self) -> Incomplete: ... - def leaf(self, index: Incomplete) -> Incomplete: ... - def leaves(self) -> Incomplete: ... - def edges(self) -> Incomplete: ... - def iteredges(self) -> Incomplete: ... - __iter__ = iteredges - def num_edges(self) -> Incomplete: ... - def select(self, **restrictions: Incomplete) -> Incomplete: ... - def insert_with_backpointer( - self, - new_edge: Incomplete, - previous_edge: Incomplete, - child_edge: Incomplete, - ) -> Incomplete: ... - def insert( - self, edge: Incomplete, *child_pointer_lists: Incomplete - ) -> Incomplete: ... - def parses( - self, root: Incomplete, tree_class: Incomplete = ... - ) -> Generator[Incomplete, Incomplete, None]: ... - def trees( - self, - edge: Incomplete, - tree_class: Incomplete = ..., - complete: bool = False, - ) -> Incomplete: ... - def child_pointer_lists(self, edge: Incomplete) -> Incomplete: ... - def pretty_format_edge( - self, edge: Incomplete, width: Incomplete | None = None - ) -> Incomplete: ... - def pretty_format_leaves( - self, width: Incomplete | None = None - ) -> Incomplete: ... - def pretty_format(self, width: Incomplete | None = None) -> Incomplete: ... - def dot_digraph(self) -> Incomplete: ... - -class ChartRuleI: - def apply( - self, chart: Incomplete, grammar: Incomplete, *edges: Incomplete - ) -> None: ... - def apply_everywhere( - self, chart: Incomplete, grammar: Incomplete - ) -> None: ... - -class AbstractChartRule(ChartRuleI): - def apply( - self, chart: Incomplete, grammar: Incomplete, *edges: Incomplete - ) -> None: ... - def apply_everywhere( - self, chart: Incomplete, grammar: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... - -class FundamentalRule(AbstractChartRule): - NUM_EDGES: int - - def apply( - self, - chart: Incomplete, - grammar: Incomplete, - left_edge: Incomplete, - right_edge: Incomplete, - ) -> Generator[Incomplete, None, None]: ... - -class SingleEdgeFundamentalRule(FundamentalRule): - NUM_EDGES: int - - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... - -class LeafInitRule(AbstractChartRule): - NUM_EDGES: int - - def apply( - self, chart: Incomplete, grammar: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class TopDownInitRule(AbstractChartRule): - NUM_EDGES: int - - def apply( - self, chart: Incomplete, grammar: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class TopDownPredictRule(AbstractChartRule): - NUM_EDGES: int - - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class CachedTopDownPredictRule(TopDownPredictRule): - def __init__(self) -> None: ... - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class BottomUpPredictRule(AbstractChartRule): - NUM_EDGES: int - - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class BottomUpPredictCombineRule(BottomUpPredictRule): - NUM_EDGES: int - - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class EmptyPredictRule(AbstractChartRule): - NUM_EDGES: int - - def apply( - self, chart: Incomplete, grammar: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class FilteredSingleEdgeFundamentalRule(SingleEdgeFundamentalRule): ... - -class FilteredBottomUpPredictCombineRule(BottomUpPredictCombineRule): - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -TD_STRATEGY: Incomplete -BU_STRATEGY: Incomplete -BU_LC_STRATEGY: Incomplete -LC_STRATEGY: Incomplete - -class ChartParser(ParserI): - def __init__( - self, - grammar: Incomplete, - strategy: Incomplete = ..., - trace: int = 0, - trace_chart_width: int = 50, - use_agenda: bool = True, - chart_class: Incomplete = ..., - ) -> None: ... - def grammar(self) -> Incomplete: ... - def chart_parse( - self, tokens: Incomplete, trace: Incomplete | None = None - ) -> Incomplete: ... - def parse( - self, tokens: Incomplete, tree_class: Incomplete = ... - ) -> Incomplete: ... - -class TopDownChartParser(ChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class BottomUpChartParser(ChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class BottomUpLeftCornerChartParser(ChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class LeftCornerChartParser(ChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class SteppingChartParser(ChartParser): - def __init__( - self, grammar: Incomplete, strategy: Incomplete = [], trace: int = 0 - ) -> None: ... - def initialize(self, tokens: Incomplete) -> None: ... - def step(self) -> Generator[Incomplete, None, None]: ... - def strategy(self) -> Incomplete: ... - def grammar(self) -> Incomplete: ... - def chart(self) -> Incomplete: ... - def current_chartrule(self) -> Incomplete: ... - def parses(self, tree_class: Incomplete = ...) -> Incomplete: ... - def set_strategy(self, strategy: Incomplete) -> None: ... - def set_grammar(self, grammar: Incomplete) -> None: ... - def set_chart(self, chart: Incomplete) -> None: ... - def parse( - self, tokens: Incomplete, tree_class: Incomplete = ... - ) -> Incomplete: ... - -def demo_grammar() -> Incomplete: ... -def demo( - choice: Incomplete | None = None, - print_times: bool = True, - print_grammar: bool = False, - print_trees: bool = True, - trace: int = 2, - sent: str = "I saw John with a dog with my cookie", - numparses: int = 5, -) -> Incomplete: ... diff --git a/stubs/nltk/parse/corenlp.pyi b/stubs/nltk/parse/corenlp.pyi deleted file mode 100644 index 2a83d13..0000000 --- a/stubs/nltk/parse/corenlp.pyi +++ /dev/null @@ -1,103 +0,0 @@ -import types -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.internals import config_java as config_java -from nltk.internals import find_jar_iter as find_jar_iter -from nltk.internals import java as java -from nltk.parse.api import ParserI as ParserI -from nltk.parse.dependencygraph import DependencyGraph as DependencyGraph -from nltk.tag.api import TaggerI as TaggerI -from nltk.tokenize.api import TokenizerI as TokenizerI -from nltk.tree import Tree as Tree - -class CoreNLPServerError(EnvironmentError): ... - -def try_port(port: int = 0) -> Incomplete: ... - -class CoreNLPServer: - url: Incomplete - verbose: Incomplete - corenlp_options: Incomplete - java_options: Incomplete - def __init__( - self, - path_to_jar: Incomplete | None = None, - path_to_models_jar: Incomplete | None = None, - verbose: bool = False, - java_options: Incomplete | None = None, - corenlp_options: Incomplete | None = None, - port: Incomplete | None = None, - ) -> None: ... - popen: Incomplete - def start( - self, stdout: str = "devnull", stderr: str = "devnull" - ) -> None: ... - def stop(self) -> None: ... - def __enter__(self) -> Incomplete: ... - def __exit__( - self, - exc_type: type[BaseException] | None, - exc_val: BaseException | None, - exc_tb: types.TracebackType | None, - ) -> Incomplete: ... - -class GenericCoreNLPParser(ParserI, TokenizerI, TaggerI): - url: Incomplete - encoding: Incomplete - tagtype: Incomplete - strict_json: Incomplete - session: Incomplete - def __init__( - self, - url: str = "http://localhost:9000", - encoding: str = "utf8", - tagtype: Incomplete | None = None, - strict_json: bool = True, - ) -> None: ... - def parse_sents( - self, sentences: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> Incomplete: ... - def raw_parse( - self, - sentence: Incomplete, - properties: Incomplete | None = None, - *args: Incomplete, - **kwargs: Incomplete, - ) -> Incomplete: ... - def api_call( - self, - data: Incomplete, - properties: Incomplete | None = None, - timeout: int = 60, - ) -> Incomplete: ... - def raw_parse_sents( - self, - sentences: Incomplete, - verbose: bool = False, - properties: Incomplete | None = None, - *args: Incomplete, - **kwargs: Incomplete, - ) -> Generator[Incomplete, None, None]: ... - def parse_text( - self, text: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> Generator[Incomplete, None, None]: ... - def tokenize( - self, text: Incomplete, properties: Incomplete | None = None - ) -> Generator[Incomplete, None, None]: ... - def tag_sents(self, sentences: Incomplete) -> Incomplete: ... - def tag(self, sentence: str) -> list[tuple[str, str]]: ... - def raw_tag_sents( - self, sentences: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class CoreNLPParser(GenericCoreNLPParser): - parser_annotator: str - def make_tree(self, result: Incomplete) -> Incomplete: ... - -class CoreNLPDependencyParser(GenericCoreNLPParser): - parser_annotator: str - def make_tree(self, result: Incomplete) -> Incomplete: ... - -def transform(sentence: Incomplete) -> Generator[Incomplete, None, None]: ... diff --git a/stubs/nltk/parse/dependencygraph.pyi b/stubs/nltk/parse/dependencygraph.pyi deleted file mode 100644 index 2d3677e..0000000 --- a/stubs/nltk/parse/dependencygraph.pyi +++ /dev/null @@ -1,64 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.internals import find_binary as find_binary -from nltk.tree import Tree as Tree - -class DependencyGraph: - nodes: Incomplete - root: Incomplete - def __init__( - self, - tree_str: Incomplete | None = None, - cell_extractor: Incomplete | None = None, - zero_based: bool = False, - cell_separator: Incomplete | None = None, - top_relation_label: str = "ROOT", - ) -> None: ... - def remove_by_address(self, address: Incomplete) -> None: ... - def redirect_arcs( - self, originals: Incomplete, redirect: Incomplete - ) -> None: ... - def add_arc( - self, head_address: Incomplete, mod_address: Incomplete - ) -> None: ... - def connect_graph(self) -> None: ... - def get_by_address(self, node_address: Incomplete) -> Incomplete: ... - def contains_address(self, node_address: Incomplete) -> Incomplete: ... - def to_dot(self) -> Incomplete: ... - @staticmethod - def load( - filename: Incomplete, - zero_based: bool = False, - cell_separator: Incomplete | None = None, - top_relation_label: str = "ROOT", - ) -> Incomplete: ... - def left_children(self, node_index: Incomplete) -> Incomplete: ... - def right_children(self, node_index: Incomplete) -> Incomplete: ... - def add_node(self, node: Incomplete) -> None: ... - def tree(self) -> Incomplete: ... - def triples( - self, node: Incomplete | None = None - ) -> Generator[Incomplete, Incomplete, None]: ... - def contains_cycle(self) -> Incomplete: ... - def get_cycle_path( - self, curr_node: Incomplete, goal_node_index: Incomplete - ) -> Incomplete: ... - def to_conll(self, style: Incomplete) -> Incomplete: ... - nx_labels: Incomplete - def nx_graph(self) -> Incomplete: ... - -def dot2img(dot_string: Incomplete, t: str = "svg") -> Incomplete: ... - -class DependencyGraphError(Exception): ... - -def demo() -> None: ... -def malt_demo(nx: bool = False) -> None: ... -def conll_demo() -> None: ... -def conll_file_demo() -> None: ... -def cycle_finding_demo() -> None: ... - -treebank_data: str -conll_data1: str -conll_data2: str diff --git a/stubs/nltk/parse/earleychart.pyi b/stubs/nltk/parse/earleychart.pyi deleted file mode 100644 index 4ba9426..0000000 --- a/stubs/nltk/parse/earleychart.pyi +++ /dev/null @@ -1,176 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.parse.chart import ( - BottomUpPredictCombineRule as BottomUpPredictCombineRule, -) -from nltk.parse.chart import BottomUpPredictRule as BottomUpPredictRule -from nltk.parse.chart import ( - CachedTopDownPredictRule as CachedTopDownPredictRule, -) -from nltk.parse.chart import Chart as Chart -from nltk.parse.chart import ChartParser as ChartParser -from nltk.parse.chart import EdgeI as EdgeI -from nltk.parse.chart import EmptyPredictRule as EmptyPredictRule -from nltk.parse.chart import ( - FilteredBottomUpPredictCombineRule as FilteredBottomUpPredictCombineRule, -) -from nltk.parse.chart import ( - FilteredSingleEdgeFundamentalRule as FilteredSingleEdgeFundamentalRule, -) -from nltk.parse.chart import LeafEdge as LeafEdge -from nltk.parse.chart import LeafInitRule as LeafInitRule -from nltk.parse.chart import ( - SingleEdgeFundamentalRule as SingleEdgeFundamentalRule, -) -from nltk.parse.chart import TopDownInitRule as TopDownInitRule -from nltk.parse.featurechart import ( - FeatureBottomUpPredictCombineRule as FeatureBottomUpPredictCombineRule, -) -from nltk.parse.featurechart import ( - FeatureBottomUpPredictRule as FeatureBottomUpPredictRule, -) -from nltk.parse.featurechart import FeatureChart as FeatureChart -from nltk.parse.featurechart import FeatureChartParser as FeatureChartParser -from nltk.parse.featurechart import ( - FeatureEmptyPredictRule as FeatureEmptyPredictRule, -) -from nltk.parse.featurechart import ( - FeatureSingleEdgeFundamentalRule as FeatureSingleEdgeFundamentalRule, -) -from nltk.parse.featurechart import ( - FeatureTopDownInitRule as FeatureTopDownInitRule, -) -from nltk.parse.featurechart import ( - FeatureTopDownPredictRule as FeatureTopDownPredictRule, -) - -class IncrementalChart(Chart): - def initialize(self) -> None: ... - def edges(self) -> Incomplete: ... - def iteredges(self) -> Incomplete: ... - def select( - self, end: Incomplete, **restrictions: Incomplete - ) -> Incomplete: ... - -class FeatureIncrementalChart(IncrementalChart, FeatureChart): - def select( - self, end: Incomplete, **restrictions: Incomplete - ) -> Incomplete: ... - -class CompleteFundamentalRule(SingleEdgeFundamentalRule): ... - -class CompleterRule(CompleteFundamentalRule): - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... - -class ScannerRule(CompleteFundamentalRule): - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... - -class PredictorRule(CachedTopDownPredictRule): ... - -class FilteredCompleteFundamentalRule(FilteredSingleEdgeFundamentalRule): - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... - -class FeatureCompleteFundamentalRule(FeatureSingleEdgeFundamentalRule): ... -class FeatureCompleterRule(CompleterRule): ... -class FeatureScannerRule(ScannerRule): ... -class FeaturePredictorRule(FeatureTopDownPredictRule): ... - -EARLEY_STRATEGY: Incomplete -TD_INCREMENTAL_STRATEGY: Incomplete -BU_INCREMENTAL_STRATEGY: Incomplete -BU_LC_INCREMENTAL_STRATEGY: Incomplete -LC_INCREMENTAL_STRATEGY: Incomplete - -class IncrementalChartParser(ChartParser): - def __init__( - self, - grammar: Incomplete, - strategy: Incomplete = ..., - trace: int = 0, - trace_chart_width: int = 50, - chart_class: Incomplete = ..., - ) -> None: ... - def chart_parse( - self, tokens: Incomplete, trace: Incomplete | None = None - ) -> Incomplete: ... - -class EarleyChartParser(IncrementalChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class IncrementalTopDownChartParser(IncrementalChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class IncrementalBottomUpChartParser(IncrementalChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class IncrementalBottomUpLeftCornerChartParser(IncrementalChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class IncrementalLeftCornerChartParser(IncrementalChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -EARLEY_FEATURE_STRATEGY: Incomplete -TD_INCREMENTAL_FEATURE_STRATEGY: Incomplete -BU_INCREMENTAL_FEATURE_STRATEGY: Incomplete -BU_LC_INCREMENTAL_FEATURE_STRATEGY: Incomplete - -class FeatureIncrementalChartParser( - IncrementalChartParser, FeatureChartParser -): - def __init__( - self, - grammar: Incomplete, - strategy: Incomplete = ..., - trace_chart_width: int = 20, - chart_class: Incomplete = ..., - **parser_args: Incomplete, - ) -> None: ... - -class FeatureEarleyChartParser(FeatureIncrementalChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class FeatureIncrementalTopDownChartParser(FeatureIncrementalChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class FeatureIncrementalBottomUpChartParser(FeatureIncrementalChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class FeatureIncrementalBottomUpLeftCornerChartParser( - FeatureIncrementalChartParser -): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -def demo( - print_times: bool = True, - print_grammar: bool = False, - print_trees: bool = True, - trace: int = 2, - sent: str = "I saw John with a dog with my cookie", - numparses: int = 5, -) -> None: ... diff --git a/stubs/nltk/parse/evaluate.pyi b/stubs/nltk/parse/evaluate.pyi deleted file mode 100644 index ff83300..0000000 --- a/stubs/nltk/parse/evaluate.pyi +++ /dev/null @@ -1,7 +0,0 @@ -from _typeshed import Incomplete - -class DependencyEvaluator: - def __init__( - self, parsed_sents: Incomplete, gold_sents: Incomplete - ) -> None: ... - def eval(self) -> Incomplete: ... diff --git a/stubs/nltk/parse/featurechart.pyi b/stubs/nltk/parse/featurechart.pyi deleted file mode 100644 index fa2b76b..0000000 --- a/stubs/nltk/parse/featurechart.pyi +++ /dev/null @@ -1,146 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.featstruct import TYPE as TYPE -from nltk.featstruct import FeatStruct as FeatStruct -from nltk.featstruct import find_variables as find_variables -from nltk.featstruct import unify as unify -from nltk.grammar import CFG as CFG -from nltk.grammar import FeatStructNonterminal as FeatStructNonterminal -from nltk.grammar import Nonterminal as Nonterminal -from nltk.grammar import Production as Production -from nltk.grammar import is_nonterminal as is_nonterminal -from nltk.grammar import is_terminal as is_terminal -from nltk.parse.chart import ( - BottomUpPredictCombineRule as BottomUpPredictCombineRule, -) -from nltk.parse.chart import BottomUpPredictRule as BottomUpPredictRule -from nltk.parse.chart import ( - CachedTopDownPredictRule as CachedTopDownPredictRule, -) -from nltk.parse.chart import Chart as Chart -from nltk.parse.chart import ChartParser as ChartParser -from nltk.parse.chart import EdgeI as EdgeI -from nltk.parse.chart import EmptyPredictRule as EmptyPredictRule -from nltk.parse.chart import FundamentalRule as FundamentalRule -from nltk.parse.chart import LeafInitRule as LeafInitRule -from nltk.parse.chart import ( - SingleEdgeFundamentalRule as SingleEdgeFundamentalRule, -) -from nltk.parse.chart import TopDownInitRule as TopDownInitRule -from nltk.parse.chart import TreeEdge as TreeEdge -from nltk.sem import logic as logic -from nltk.tree import Tree as Tree - -class FeatureTreeEdge(TreeEdge): - def __init__( - self, - span: Incomplete, - lhs: Incomplete, - rhs: Incomplete, - dot: int = 0, - bindings: Incomplete | None = None, - ) -> None: ... - @staticmethod - def from_production( - production: Incomplete, index: Incomplete - ) -> Incomplete: ... - def move_dot_forward( - self, new_end: Incomplete, bindings: Incomplete | None = None - ) -> Incomplete: ... - def next_with_bindings(self) -> Incomplete: ... - def bindings(self) -> Incomplete: ... - def variables(self) -> Incomplete: ... - -class FeatureChart(Chart): - def select(self, **restrictions: Incomplete) -> Incomplete: ... - def parses( - self, start: Incomplete, tree_class: Incomplete = ... - ) -> Generator[Incomplete, Incomplete, None]: ... - -class FeatureFundamentalRule(FundamentalRule): - def apply( - self, - chart: Incomplete, - grammar: Incomplete, - left_edge: Incomplete, - right_edge: Incomplete, - ) -> Generator[Incomplete, None, None]: ... - -class FeatureSingleEdgeFundamentalRule(SingleEdgeFundamentalRule): ... - -class FeatureTopDownInitRule(TopDownInitRule): - def apply( - self, chart: Incomplete, grammar: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class FeatureTopDownPredictRule(CachedTopDownPredictRule): - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class FeatureBottomUpPredictRule(BottomUpPredictRule): - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class FeatureBottomUpPredictCombineRule(BottomUpPredictCombineRule): - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class FeatureEmptyPredictRule(EmptyPredictRule): - def apply( - self, chart: Incomplete, grammar: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -TD_FEATURE_STRATEGY: Incomplete -BU_FEATURE_STRATEGY: Incomplete -BU_LC_FEATURE_STRATEGY: Incomplete - -class FeatureChartParser(ChartParser): - def __init__( - self, - grammar: Incomplete, - strategy: Incomplete = ..., - trace_chart_width: int = 20, - chart_class: Incomplete = ..., - **parser_args: Incomplete, - ) -> None: ... - -class FeatureTopDownChartParser(FeatureChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class FeatureBottomUpChartParser(FeatureChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class FeatureBottomUpLeftCornerChartParser(FeatureChartParser): - def __init__( - self, grammar: Incomplete, **parser_args: Incomplete - ) -> None: ... - -class InstantiateVarsChart(FeatureChart): - def __init__(self, tokens: Incomplete) -> None: ... - def initialize(self) -> None: ... - def insert( - self, edge: Incomplete, child_pointer_list: Incomplete - ) -> Incomplete: ... - def instantiate_edge(self, edge: Incomplete) -> None: ... - def inst_vars(self, edge: Incomplete) -> Incomplete: ... - -def demo_grammar() -> Incomplete: ... -def demo( - print_times: bool = True, - print_grammar: bool = True, - print_trees: bool = True, - print_sentence: bool = True, - trace: int = 1, - parser: Incomplete = ..., - sent: str = "I saw John with a dog with my cookie", -) -> None: ... -def run_profile() -> None: ... diff --git a/stubs/nltk/parse/generate.pyi b/stubs/nltk/parse/generate.pyi deleted file mode 100644 index 072de2e..0000000 --- a/stubs/nltk/parse/generate.pyi +++ /dev/null @@ -1,14 +0,0 @@ -from _typeshed import Incomplete - -from nltk.grammar import Nonterminal as Nonterminal - -def generate( - grammar: Incomplete, - start: Incomplete | None = None, - depth: Incomplete | None = None, - n: Incomplete | None = None, -) -> Incomplete: ... - -demo_grammar: str - -def demo(N: int = 23) -> None: ... diff --git a/stubs/nltk/parse/malt.pyi b/stubs/nltk/parse/malt.pyi deleted file mode 100644 index 9171f1f..0000000 --- a/stubs/nltk/parse/malt.pyi +++ /dev/null @@ -1,57 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.data import ZipFilePathPointer as ZipFilePathPointer -from nltk.internals import ( - find_dir as find_dir, -) -from nltk.internals import ( - find_file as find_file, -) -from nltk.internals import ( - find_jars_within_path as find_jars_within_path, -) -from nltk.parse.api import ParserI as ParserI -from nltk.parse.dependencygraph import DependencyGraph as DependencyGraph -from nltk.parse.util import taggedsents_to_conll as taggedsents_to_conll - -def malt_regex_tagger() -> Incomplete: ... -def find_maltparser(parser_dirname: Incomplete) -> Incomplete: ... -def find_malt_model(model_filename: Incomplete) -> Incomplete: ... - -class MaltParser(ParserI): - malt_jars: Incomplete - additional_java_args: Incomplete - model: Incomplete - working_dir: Incomplete - tagger: Incomplete - def __init__( - self, - parser_dirname: str = "", - model_filename: Incomplete | None = None, - tagger: Incomplete | None = None, - additional_java_args: Incomplete | None = None, - ) -> None: ... - def parse_tagged_sents( - self, - sentences: Incomplete, - verbose: bool = False, - top_relation_label: str = "null", - ) -> Generator[Incomplete, None, None]: ... - def parse_sents( - self, - sentences: Incomplete, - verbose: bool = False, - top_relation_label: str = "null", - ) -> Incomplete: ... - def generate_malt_command( - self, - inputfilename: Incomplete, - outputfilename: Incomplete | None = None, - mode: Incomplete | None = None, - ) -> Incomplete: ... - def train(self, depgraphs: Incomplete, verbose: bool = False) -> None: ... - def train_from_file( - self, conll_file: Incomplete, verbose: bool = False - ) -> Incomplete: ... diff --git a/stubs/nltk/parse/nonprojectivedependencyparser.pyi b/stubs/nltk/parse/nonprojectivedependencyparser.pyi deleted file mode 100644 index ffd2533..0000000 --- a/stubs/nltk/parse/nonprojectivedependencyparser.pyi +++ /dev/null @@ -1,64 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.parse.dependencygraph import DependencyGraph as DependencyGraph - -logger: Incomplete - -class DependencyScorerI: - def __init__(self) -> None: ... - def train(self, graphs: Incomplete) -> None: ... - def score(self, graph: Incomplete) -> None: ... - -class NaiveBayesDependencyScorer(DependencyScorerI): - def __init__(self) -> None: ... - classifier: Incomplete - def train(self, graphs: Incomplete) -> None: ... - def score(self, graph: Incomplete) -> Incomplete: ... - -class DemoScorer(DependencyScorerI): - def train(self, graphs: Incomplete) -> None: ... - def score(self, graph: Incomplete) -> Incomplete: ... - -class ProbabilisticNonprojectiveParser: - def __init__(self) -> None: ... - def train( - self, graphs: Incomplete, dependency_scorer: Incomplete - ) -> None: ... - scores: Incomplete - def initialize_edge_scores(self, graph: Incomplete) -> None: ... - def collapse_nodes( - self, - new_node: Incomplete, - cycle_path: Incomplete, - g_graph: Incomplete, - b_graph: Incomplete, - c_graph: Incomplete, - ) -> None: ... - def update_edge_scores( - self, new_node: Incomplete, cycle_path: Incomplete - ) -> None: ... - def compute_original_indexes( - self, new_indexes: Incomplete - ) -> Incomplete: ... - def compute_max_subtract_score( - self, column_index: Incomplete, cycle_indexes: Incomplete - ) -> Incomplete: ... - def best_incoming_arc(self, node_inde: Incomplete) -> Incomplete: ... - def original_best_arc(self, node_inde: Incomplete) -> Incomplete: ... - inner_nodes: Incomplete - def parse( - self, tokens: Incomplete, tags: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class NonprojectiveDependencyParser: - def __init__(self, dependency_grammar: Incomplete) -> None: ... - def parse( - self, tokens: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -def demo() -> None: ... -def hall_demo() -> None: ... -def nonprojective_conll_parse_demo() -> None: ... -def rule_based_demo() -> None: ... diff --git a/stubs/nltk/parse/pchart.pyi b/stubs/nltk/parse/pchart.pyi deleted file mode 100644 index 5e59189..0000000 --- a/stubs/nltk/parse/pchart.pyi +++ /dev/null @@ -1,94 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.grammar import PCFG as PCFG -from nltk.grammar import Nonterminal as Nonterminal -from nltk.parse.api import ParserI as ParserI -from nltk.parse.chart import ( - AbstractChartRule as AbstractChartRule, -) -from nltk.parse.chart import ( - Chart as Chart, -) -from nltk.parse.chart import ( - LeafEdge as LeafEdge, -) -from nltk.parse.chart import ( - TreeEdge as TreeEdge, -) -from nltk.tree import ProbabilisticTree as ProbabilisticTree -from nltk.tree import Tree as Tree - -class ProbabilisticLeafEdge(LeafEdge): - def prob(self) -> Incomplete: ... - -class ProbabilisticTreeEdge(TreeEdge): - def __init__( - self, prob: Incomplete, *args: Incomplete, **kwargs: Incomplete - ) -> None: ... - def prob(self) -> Incomplete: ... - @staticmethod - def from_production( - production: Incomplete, index: Incomplete, p: Incomplete - ) -> Incomplete: ... - -class ProbabilisticBottomUpInitRule(AbstractChartRule): - NUM_EDGES: int - def apply( - self, chart: Incomplete, grammar: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class ProbabilisticBottomUpPredictRule(AbstractChartRule): - NUM_EDGES: int - def apply( - self, chart: Incomplete, grammar: Incomplete, edge: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class ProbabilisticFundamentalRule(AbstractChartRule): - NUM_EDGES: int - def apply( - self, - chart: Incomplete, - grammar: Incomplete, - left_edge: Incomplete, - right_edge: Incomplete, - ) -> Generator[Incomplete, None, None]: ... - -class SingleEdgeProbabilisticFundamentalRule(AbstractChartRule): - NUM_EDGES: int - def apply( - self, chart: Incomplete, grammar: Incomplete, edge1: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... - -class BottomUpProbabilisticChartParser(ParserI): - beam_size: Incomplete - def __init__( - self, grammar: Incomplete, beam_size: int = 0, trace: int = 0 - ) -> None: ... - def grammar(self) -> Incomplete: ... - def trace(self, trace: int = 2) -> None: ... - def parse(self, tokens: Incomplete) -> Incomplete: ... - def sort_queue(self, queue: Incomplete, chart: Incomplete) -> None: ... - -class InsideChartParser(BottomUpProbabilisticChartParser): - def sort_queue( - self, queue: Incomplete, chart: Incomplete - ) -> Incomplete: ... - -class RandomChartParser(BottomUpProbabilisticChartParser): - def sort_queue(self, queue: Incomplete, chart: Incomplete) -> None: ... - -class UnsortedChartParser(BottomUpProbabilisticChartParser): - def sort_queue(self, queue: Incomplete, chart: Incomplete) -> None: ... - -class LongestChartParser(BottomUpProbabilisticChartParser): - def sort_queue( - self, queue: Incomplete, chart: Incomplete - ) -> Incomplete: ... - -def demo( - choice: Incomplete | None = None, - draw_parses: Incomplete | None = None, - print_parses: Incomplete | None = None, -) -> Incomplete: ... diff --git a/stubs/nltk/parse/projectivedependencyparser.pyi b/stubs/nltk/parse/projectivedependencyparser.pyi deleted file mode 100644 index 3b332f9..0000000 --- a/stubs/nltk/parse/projectivedependencyparser.pyi +++ /dev/null @@ -1,57 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.grammar import ( - DependencyGrammar as DependencyGrammar, -) -from nltk.grammar import ( - DependencyProduction as DependencyProduction, -) -from nltk.grammar import ( - ProbabilisticDependencyGrammar as ProbabilisticDependencyGrammar, -) -from nltk.internals import raise_unorderable_types as raise_unorderable_types -from nltk.parse.dependencygraph import DependencyGraph as DependencyGraph - -class DependencySpan: - def __init__( - self, - start_index: Incomplete, - end_index: Incomplete, - head_index: Incomplete, - arcs: Incomplete, - tags: Incomplete, - ) -> None: ... - def head_index(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class ChartCell: - def __init__(self, x: Incomplete, y: Incomplete) -> None: ... - def add(self, span: Incomplete) -> None: ... - -class ProjectiveDependencyParser: - def __init__(self, dependency_grammar: Incomplete) -> None: ... - def parse( - self, tokens: Incomplete - ) -> Generator[Incomplete, None, None]: ... - def concatenate( - self, span1: Incomplete, span2: Incomplete - ) -> Incomplete: ... - -class ProbabilisticProjectiveDependencyParser: - def __init__(self) -> None: ... - def parse(self, tokens: Incomplete) -> Incomplete: ... - def concatenate( - self, span1: Incomplete, span2: Incomplete - ) -> Incomplete: ... - def train(self, graphs: Incomplete) -> None: ... - def compute_prob(self, dg: Incomplete) -> Incomplete: ... - -def demo() -> None: ... -def projective_rule_parse_demo() -> None: ... -def arity_parse_demo() -> None: ... -def projective_prob_parse_demo() -> None: ... diff --git a/stubs/nltk/parse/recursivedescent.pyi b/stubs/nltk/parse/recursivedescent.pyi deleted file mode 100644 index c52caa5..0000000 --- a/stubs/nltk/parse/recursivedescent.pyi +++ /dev/null @@ -1,32 +0,0 @@ -from _typeshed import Incomplete - -from nltk.grammar import Nonterminal as Nonterminal -from nltk.parse.api import ParserI as ParserI -from nltk.tree import ImmutableTree as ImmutableTree -from nltk.tree import Tree as Tree - -class RecursiveDescentParser(ParserI): - def __init__(self, grammar: Incomplete, trace: int = 0) -> None: ... - def grammar(self) -> Incomplete: ... - def parse(self, tokens: Incomplete) -> Incomplete: ... - def trace(self, trace: int = 2) -> None: ... - -class SteppingRecursiveDescentParser(RecursiveDescentParser): - def __init__(self, grammar: Incomplete, trace: int = 0) -> None: ... - def parse(self, tokens: Incomplete) -> Incomplete: ... - def initialize(self, tokens: Incomplete) -> None: ... - def remaining_text(self) -> Incomplete: ... - def frontier(self) -> Incomplete: ... - def tree(self) -> Incomplete: ... - def step(self) -> Incomplete: ... - def expand(self, production: Incomplete | None = None) -> Incomplete: ... - def match(self) -> Incomplete: ... - def backtrack(self) -> Incomplete: ... - def expandable_productions(self) -> Incomplete: ... - def untried_expandable_productions(self) -> Incomplete: ... - def untried_match(self) -> Incomplete: ... - def currently_complete(self) -> Incomplete: ... - def parses(self) -> Incomplete: ... - def set_grammar(self, grammar: Incomplete) -> None: ... - -def demo() -> None: ... diff --git a/stubs/nltk/parse/shiftreduce.pyi b/stubs/nltk/parse/shiftreduce.pyi deleted file mode 100644 index e86f68b..0000000 --- a/stubs/nltk/parse/shiftreduce.pyi +++ /dev/null @@ -1,31 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.grammar import Nonterminal as Nonterminal -from nltk.parse.api import ParserI as ParserI -from nltk.tree import Tree as Tree - -class ShiftReduceParser(ParserI): - def __init__(self, grammar: Incomplete, trace: int = 0) -> None: ... - def grammar(self) -> Incomplete: ... - def parse( - self, tokens: Incomplete - ) -> Generator[Incomplete, None, None]: ... - def trace(self, trace: int = 2) -> None: ... - -class SteppingShiftReduceParser(ShiftReduceParser): - def __init__(self, grammar: Incomplete, trace: int = 0) -> None: ... - def parse(self, tokens: Incomplete) -> Incomplete: ... - def stack(self) -> Incomplete: ... - def remaining_text(self) -> Incomplete: ... - def initialize(self, tokens: Incomplete) -> None: ... - def step(self) -> Incomplete: ... - def shift(self) -> Incomplete: ... - def reduce(self, production: Incomplete | None = None) -> Incomplete: ... - def undo(self) -> Incomplete: ... - def reducible_productions(self) -> Incomplete: ... - def parses(self) -> Generator[Incomplete, None, None]: ... - def set_grammar(self, grammar: Incomplete) -> None: ... - -def demo() -> None: ... diff --git a/stubs/nltk/parse/stanford.pyi b/stubs/nltk/parse/stanford.pyi deleted file mode 100644 index 4ed0fd3..0000000 --- a/stubs/nltk/parse/stanford.pyi +++ /dev/null @@ -1,59 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import ( - config_java as config_java, -) -from nltk.internals import ( - find_jar_iter as find_jar_iter, -) -from nltk.internals import ( - find_jars_within_path as find_jars_within_path, -) -from nltk.internals import ( - java as java, -) -from nltk.parse.api import ParserI as ParserI -from nltk.parse.dependencygraph import DependencyGraph as DependencyGraph -from nltk.tree import Tree as Tree - -class GenericStanfordParser(ParserI): - model_path: Incomplete - corenlp_options: Incomplete - java_options: Incomplete - def __init__( - self, - path_to_jar: Incomplete | None = None, - path_to_models_jar: Incomplete | None = None, - model_path: str = "edu/stanford/nlp/models/lexparser/englishPCFG.ser.gz", - encoding: str = "utf8", - verbose: bool = False, - java_options: str = "-mx4g", - corenlp_options: str = "", - ) -> None: ... - def parse_sents( - self, sentences: Incomplete, verbose: bool = False - ) -> Incomplete: ... - def raw_parse( - self, sentence: Incomplete, verbose: bool = False - ) -> Incomplete: ... - def raw_parse_sents( - self, sentences: Incomplete, verbose: bool = False - ) -> Incomplete: ... - def tagged_parse( - self, sentence: Incomplete, verbose: bool = False - ) -> Incomplete: ... - def tagged_parse_sents( - self, sentences: Incomplete, verbose: bool = False - ) -> Incomplete: ... - -class StanfordParser(GenericStanfordParser): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class StanfordDependencyParser(GenericStanfordParser): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class StanfordNeuralDependencyParser(GenericStanfordParser): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def tagged_parse_sents( - self, sentences: Incomplete, verbose: bool = False - ) -> None: ... diff --git a/stubs/nltk/parse/transitionparser.pyi b/stubs/nltk/parse/transitionparser.pyi deleted file mode 100644 index 2fce072..0000000 --- a/stubs/nltk/parse/transitionparser.pyi +++ /dev/null @@ -1,49 +0,0 @@ -from _typeshed import Incomplete - -from nltk.parse import ( - DependencyEvaluator as DependencyEvaluator, -) -from nltk.parse import ( - DependencyGraph as DependencyGraph, -) -from nltk.parse import ( - ParserI as ParserI, -) - -class Configuration: - stack: Incomplete - buffer: Incomplete - arcs: Incomplete - def __init__(self, dep_graph: Incomplete) -> None: ... - def extract_features(self) -> Incomplete: ... - -class Transition: - LEFT_ARC: str - RIGHT_ARC: str - SHIFT: str - REDUCE: str - def __init__(self, alg_option: Incomplete) -> None: ... - def left_arc( - self, conf: Incomplete, relation: Incomplete - ) -> Incomplete: ... - def right_arc( - self, conf: Incomplete, relation: Incomplete - ) -> Incomplete: ... - def reduce(self, conf: Incomplete) -> Incomplete: ... - def shift(self, conf: Incomplete) -> Incomplete: ... - -class TransitionParser(ParserI): - ARC_STANDARD: str - ARC_EAGER: str - def __init__(self, algorithm: Incomplete) -> None: ... - def train( - self, - depgraphs: Incomplete, - modelfile: Incomplete, - verbose: bool = True, - ) -> None: ... - def parse( - self, depgraphs: Incomplete, modelFile: Incomplete - ) -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/parse/util.pyi b/stubs/nltk/parse/util.pyi deleted file mode 100644 index 191fb25..0000000 --- a/stubs/nltk/parse/util.pyi +++ /dev/null @@ -1,57 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.data import load as load -from nltk.grammar import ( - CFG as CFG, -) -from nltk.grammar import ( - PCFG as PCFG, -) -from nltk.grammar import ( - FeatureGrammar as FeatureGrammar, -) -from nltk.parse.chart import Chart as Chart -from nltk.parse.chart import ChartParser as ChartParser -from nltk.parse.featurechart import ( - FeatureChart as FeatureChart, -) -from nltk.parse.featurechart import ( - FeatureChartParser as FeatureChartParser, -) -from nltk.parse.pchart import InsideChartParser as InsideChartParser - -def load_parser( - grammar_url: Incomplete, - trace: int = 0, - parser: Incomplete | None = None, - chart_class: Incomplete | None = None, - beam_size: int = 0, - **load_args: Incomplete, -) -> Incomplete: ... -def taggedsent_to_conll( - sentence: Incomplete, -) -> Generator[Incomplete, None, None]: ... -def taggedsents_to_conll( - sentences: Incomplete, -) -> Generator[Incomplete, Incomplete, None]: ... - -class TestGrammar: - test_grammar: Incomplete - cp: Incomplete - suite: Incomplete - def __init__( - self, - grammar: Incomplete, - suite: Incomplete, - accept: Incomplete | None = None, - reject: Incomplete | None = None, - ) -> None: ... - def run(self, show_trees: bool = False) -> None: ... - -def extract_test_sentences( - string: Incomplete, - comment_chars: str = "#%;", - encoding: Incomplete | None = None, -) -> Incomplete: ... diff --git a/stubs/nltk/parse/viterbi.pyi b/stubs/nltk/parse/viterbi.pyi deleted file mode 100644 index ff61d07..0000000 --- a/stubs/nltk/parse/viterbi.pyi +++ /dev/null @@ -1,17 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.parse.api import ParserI as ParserI -from nltk.tree import ProbabilisticTree as ProbabilisticTree -from nltk.tree import Tree as Tree - -class ViterbiParser(ParserI): - def __init__(self, grammar: Incomplete, trace: int = 0) -> None: ... - def grammar(self) -> Incomplete: ... - def trace(self, trace: int = 2) -> None: ... - def parse( - self, tokens: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -def demo() -> Incomplete: ... diff --git a/stubs/nltk/probability.pyi b/stubs/nltk/probability.pyi deleted file mode 100644 index c2a2192..0000000 --- a/stubs/nltk/probability.pyi +++ /dev/null @@ -1,302 +0,0 @@ -from abc import ABCMeta, abstractmethod -from typing import ( - Any, - Dict, - Iterator, - List, - Optional, - Tuple, - Union, -) - -from _typeshed import Incomplete -from matplotlib.axes._axes import Axes # type: ignore[import-not-found] -from numpy import float64 - -__all__ = [ - "ConditionalFreqDist", - "ConditionalProbDist", - "ConditionalProbDistI", - "CrossValidationProbDist", - "DictionaryConditionalProbDist", - "DictionaryProbDist", - "ELEProbDist", - "FreqDist", - "SimpleGoodTuringProbDist", - "HeldoutProbDist", - "ImmutableProbabilisticMixIn", - "LaplaceProbDist", - "LidstoneProbDist", - "MLEProbDist", - "MutableProbDist", - "KneserNeyProbDist", - "ProbDistI", - "ProbabilisticMixIn", - "UniformProbDist", - "WittenBellProbDist", - "add_logs", - "log_likelihood", - "sum_logs", - "entropy", -] - -def _get_kwarg( - kwargs: Dict[str, str], key: str, default: bool -) -> Union[str, bool]: ... - -class ConditionalFreqDist: - def conditions(self) -> List[Any]: ... - def plot( - self, - *args: Incomplete, - samples: Incomplete | None = None, - title: str = "", - cumulative: bool = False, - percents: bool = False, - conditions: Incomplete | None = None, - show: bool = False, - **kwargs: Incomplete, - ) -> Axes: ... - def __init__(self, cond_samples: Incomplete | None = None) -> None: ... - def __reduce__(self) -> Incomplete: ... - def N(self) -> Incomplete: ... - def tabulate(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def __add__(self, other: Incomplete) -> Incomplete: ... - def __sub__(self, other: Incomplete) -> Incomplete: ... - def __or__(self, other: Incomplete) -> Incomplete: ... - def __and__(self, other: Incomplete) -> Incomplete: ... - def __le__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __ge__(self, other: Incomplete) -> Incomplete: ... - def __gt__(self, other: Incomplete) -> Incomplete: ... - def deepcopy(self) -> Incomplete: ... - copy = deepcopy - -class DictionaryConditionalProbDist: - def __init__( - self, probdist_dict: Dict[str, DictionaryProbDist] - ) -> None: ... - -class DictionaryProbDist: - def __init__( - self, - prob_dict: Optional[Dict[str, Union[float64, float]]] = None, - log: bool = False, - normalize: bool = False, - ) -> None: ... - def logprob(self, sample: str) -> float: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def max(self) -> Incomplete: ... - def samples(self) -> Incomplete: ... - -class FreqDist: - def B(self) -> int: ... - def N(self) -> int: ... - def __init__(self, samples: Optional[List[str]] = None) -> None: ... - def __iter__(self) -> Iterator[Union[str, Tuple[str, str]]]: ... - def __setitem__( - self, key: Union[int, str, Tuple[str, str]], val: int - ) -> None: ... - def __delitem__(self, key: Incomplete) -> None: ... - def update(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def setdefault(self, key: Incomplete, val: Incomplete) -> None: ... - def hapaxes(self) -> Incomplete: ... - def Nr( - self, r: Incomplete, bins: Incomplete | None = None - ) -> Incomplete: ... - def r_Nr(self, bins: Incomplete | None = None) -> Incomplete: ... - def freq(self, sample: Incomplete) -> Incomplete: ... - def max(self) -> Incomplete: ... - def plot( - self, - *args: Incomplete, - title: str = "", - cumulative: bool = False, - percents: bool = False, - show: bool = False, - **kwargs: Incomplete, - ) -> Incomplete: ... - def tabulate(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def copy(self) -> Incomplete: ... - def __add__(self, other: Incomplete) -> Incomplete: ... - def __sub__(self, other: Incomplete) -> Incomplete: ... - def __or__(self, other: Incomplete) -> Incomplete: ... - def __and__(self, other: Incomplete) -> Incomplete: ... - def __le__(self, other: Incomplete) -> Incomplete: ... - def __ge__(self, other: Incomplete) -> Incomplete: ... - __lt__: Incomplete - __gt__: Incomplete - def pprint( - self, maxlen: int = 10, stream: Incomplete | None = None - ) -> None: ... - def pformat(self, maxlen: int = 10) -> Incomplete: ... - -class ProbDistI(metaclass=ABCMeta): - SUM_TO_ONE: bool - @abstractmethod - def __init__(self) -> Incomplete: ... - @abstractmethod - def prob(self, sample: Incomplete) -> Incomplete: ... - def logprob(self, sample: Incomplete) -> Incomplete: ... - @abstractmethod - def max(self) -> Incomplete: ... - @abstractmethod - def samples(self) -> Incomplete: ... - def discount(self) -> Incomplete: ... - def generate(self) -> Incomplete: ... - -class UniformProbDist(ProbDistI): - def __init__(self, samples: Incomplete) -> None: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def max(self) -> Incomplete: ... - def samples(self) -> Incomplete: ... - -class RandomProbDist(ProbDistI): - def __init__(self, samples: Incomplete) -> None: ... - @classmethod - def unirand(cls: Incomplete, samples: Incomplete) -> Incomplete: ... - def max(self) -> Incomplete: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def samples(self) -> Incomplete: ... - -class MLEProbDist(ProbDistI): - def __init__( - self, freqdist: Incomplete, bins: Incomplete | None = None - ) -> None: ... - def freqdist(self) -> Incomplete: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def max(self) -> Incomplete: ... - def samples(self) -> Incomplete: ... - -class LidstoneProbDist(ProbDistI): - SUM_TO_ONE: bool - def __init__( - self, - freqdist: Incomplete, - gamma: Incomplete, - bins: Incomplete | None = None, - ) -> None: ... - def freqdist(self) -> Incomplete: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def max(self) -> Incomplete: ... - def samples(self) -> Incomplete: ... - def discount(self) -> Incomplete: ... - -class LaplaceProbDist(LidstoneProbDist): - def __init__( - self, freqdist: Incomplete, bins: Incomplete | None = None - ) -> None: ... - -class ELEProbDist(LidstoneProbDist): - def __init__( - self, freqdist: Incomplete, bins: Incomplete | None = None - ) -> None: ... - -class HeldoutProbDist(ProbDistI): - SUM_TO_ONE: bool - def __init__( - self, - base_fdist: Incomplete, - heldout_fdist: Incomplete, - bins: Incomplete | None = None, - ) -> None: ... - def base_fdist(self) -> Incomplete: ... - def heldout_fdist(self) -> Incomplete: ... - def samples(self) -> Incomplete: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def max(self) -> Incomplete: ... - def discount(self) -> None: ... - -class CrossValidationProbDist(ProbDistI): - SUM_TO_ONE: bool - def __init__(self, freqdists: Incomplete, bins: Incomplete) -> None: ... - def freqdists(self) -> Incomplete: ... - def samples(self) -> Incomplete: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def discount(self) -> None: ... - -class WittenBellProbDist(ProbDistI): - def __init__( - self, freqdist: Incomplete, bins: Incomplete | None = None - ) -> None: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def max(self) -> Incomplete: ... - def samples(self) -> Incomplete: ... - def freqdist(self) -> Incomplete: ... - def discount(self) -> None: ... - -class SimpleGoodTuringProbDist(ProbDistI): - SUM_TO_ONE: bool - def __init__( - self, freqdist: Incomplete, bins: Incomplete | None = None - ) -> None: ... - def find_best_fit(self, r: Incomplete, nr: Incomplete) -> None: ... - def smoothedNr(self, r: Incomplete) -> Incomplete: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def check(self) -> None: ... - def discount(self) -> Incomplete: ... - def max(self) -> Incomplete: ... - def samples(self) -> Incomplete: ... - def freqdist(self) -> Incomplete: ... - -class MutableProbDist(ProbDistI): - def __init__( - self, - prob_dist: Incomplete, - samples: Incomplete, - store_logs: bool = True, - ) -> None: ... - def max(self) -> Incomplete: ... - def samples(self) -> Incomplete: ... - def prob(self, sample: Incomplete) -> Incomplete: ... - def logprob(self, sample: Incomplete) -> Incomplete: ... - def update( - self, sample: Incomplete, prob: Incomplete, log: bool = True - ) -> None: ... - -class KneserNeyProbDist(ProbDistI): - def __init__( - self, - freqdist: Incomplete, - bins: Incomplete | None = None, - discount: float = 0.75, - ) -> None: ... - def prob(self, trigram: Incomplete) -> Incomplete: ... - def discount(self) -> Incomplete: ... - def set_discount(self, discount: Incomplete) -> None: ... - def samples(self) -> Incomplete: ... - def max(self) -> Incomplete: ... - -def log_likelihood( - test_pdist: Incomplete, actual_pdist: Incomplete -) -> Incomplete: ... -def entropy(pdist: Incomplete) -> Incomplete: ... - -class ConditionalProbDistI(dict, metaclass=ABCMeta): - @abstractmethod - def __init__(self) -> Incomplete: ... - def conditions(self) -> Incomplete: ... - -class ConditionalProbDist(ConditionalProbDistI): - def __init__( - self, - cfdist: Incomplete, - probdist_factory: Incomplete, - *factory_args: Incomplete, - **factory_kw_args: Incomplete, - ) -> None: ... - def __missing__(self, key: Incomplete) -> Incomplete: ... - -def add_logs(logx: Incomplete, logy: Incomplete) -> Incomplete: ... -def sum_logs(logs: Incomplete) -> Incomplete: ... - -class ProbabilisticMixIn: - def __init__(self, **kwargs: Incomplete) -> None: ... - def set_prob(self, prob: Incomplete) -> None: ... - def set_logprob(self, logprob: Incomplete) -> None: ... - def prob(self) -> Incomplete: ... - def logprob(self) -> Incomplete: ... - -class ImmutableProbabilisticMixIn(ProbabilisticMixIn): - def set_prob(self, prob: Incomplete) -> None: ... - def set_logprob(self, prob: Incomplete) -> None: ... diff --git a/stubs/nltk/sem/__init__.pyi b/stubs/nltk/sem/__init__.pyi deleted file mode 100644 index 2e6680d..0000000 --- a/stubs/nltk/sem/__init__.pyi +++ /dev/null @@ -1,74 +0,0 @@ -from nltk.sem.boxer import Boxer as Boxer -from nltk.sem.drt import DRS as DRS -from nltk.sem.drt import DrtExpression as DrtExpression -from nltk.sem.evaluate import ( - Assignment as Assignment, -) -from nltk.sem.evaluate import ( - Model as Model, -) -from nltk.sem.evaluate import ( - Undefined as Undefined, -) -from nltk.sem.evaluate import ( - Valuation as Valuation, -) -from nltk.sem.evaluate import ( - arity as arity, -) -from nltk.sem.evaluate import ( - is_rel as is_rel, -) -from nltk.sem.evaluate import ( - read_valuation as read_valuation, -) -from nltk.sem.evaluate import ( - set2rel as set2rel, -) -from nltk.sem.lfg import FStructure as FStructure -from nltk.sem.logic import ( - ApplicationExpression as ApplicationExpression, -) -from nltk.sem.logic import ( - Expression as Expression, -) -from nltk.sem.logic import ( - LogicalExpressionException as LogicalExpressionException, -) -from nltk.sem.logic import ( - Variable as Variable, -) -from nltk.sem.logic import ( - binding_ops as binding_ops, -) -from nltk.sem.logic import ( - boolean_ops as boolean_ops, -) -from nltk.sem.logic import ( - equality_preds as equality_preds, -) -from nltk.sem.logic import ( - read_logic as read_logic, -) -from nltk.sem.relextract import ( - clause as clause, -) -from nltk.sem.relextract import ( - extract_rels as extract_rels, -) -from nltk.sem.relextract import ( - rtuple as rtuple, -) -from nltk.sem.skolemize import skolemize as skolemize -from nltk.sem.util import ( - evaluate_sents as evaluate_sents, -) -from nltk.sem.util import ( - interpret_sents as interpret_sents, -) -from nltk.sem.util import ( - parse_sents as parse_sents, -) -from nltk.sem.util import ( - root_semrep as root_semrep, -) diff --git a/stubs/nltk/sem/boxer.pyi b/stubs/nltk/sem/boxer.pyi deleted file mode 100644 index 611a765..0000000 --- a/stubs/nltk/sem/boxer.pyi +++ /dev/null @@ -1,320 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import find_binary as find_binary -from nltk.sem.drt import ( - DRS as DRS, -) -from nltk.sem.drt import ( - DrtApplicationExpression as DrtApplicationExpression, -) -from nltk.sem.drt import ( - DrtEqualityExpression as DrtEqualityExpression, -) -from nltk.sem.drt import ( - DrtNegatedExpression as DrtNegatedExpression, -) -from nltk.sem.drt import ( - DrtOrExpression as DrtOrExpression, -) -from nltk.sem.drt import ( - DrtParser as DrtParser, -) -from nltk.sem.drt import ( - DrtProposition as DrtProposition, -) -from nltk.sem.drt import ( - DrtTokens as DrtTokens, -) -from nltk.sem.drt import ( - DrtVariableExpression as DrtVariableExpression, -) -from nltk.sem.logic import ( - ExpectedMoreTokensException as ExpectedMoreTokensException, -) -from nltk.sem.logic import ( - LogicalExpressionException as LogicalExpressionException, -) -from nltk.sem.logic import ( - UnexpectedTokenException as UnexpectedTokenException, -) -from nltk.sem.logic import ( - Variable as Variable, -) - -class Boxer: - def __init__( - self, - boxer_drs_interpreter: Incomplete | None = None, - elimeq: bool = False, - bin_dir: Incomplete | None = None, - verbose: bool = False, - resolve: bool = True, - ) -> None: ... - def set_bin_dir( - self, bin_dir: Incomplete, verbose: bool = False - ) -> None: ... - def interpret( - self, - input: Incomplete, - discourse_id: Incomplete | None = None, - question: bool = False, - verbose: bool = False, - ) -> Incomplete: ... - def interpret_multi( - self, - input: Incomplete, - discourse_id: Incomplete | None = None, - question: bool = False, - verbose: bool = False, - ) -> Incomplete: ... - def interpret_sents( - self, - inputs: Incomplete, - discourse_ids: Incomplete | None = None, - question: bool = False, - verbose: bool = False, - ) -> Incomplete: ... - def interpret_multi_sents( - self, - inputs: Incomplete, - discourse_ids: Incomplete | None = None, - question: bool = False, - verbose: bool = False, - ) -> Incomplete: ... - -class BoxerOutputDrsParser(DrtParser): - discourse_id: Incomplete - sentence_id_offset: Incomplete - quote_chars: Incomplete - def __init__(self, discourse_id: Incomplete | None = None) -> None: ... - def parse( - self, data: Incomplete, signature: Incomplete | None = None - ) -> Incomplete: ... - def get_all_symbols(self) -> Incomplete: ... - def handle(self, tok: Incomplete, context: Incomplete) -> Incomplete: ... - def attempt_adjuncts( - self, expression: Incomplete, context: Incomplete - ) -> Incomplete: ... - def parse_condition(self, indices: Incomplete) -> Incomplete: ... - def handle_drs(self, tok: Incomplete) -> Incomplete: ... - def handle_condition( - self, tok: Incomplete, indices: Incomplete - ) -> Incomplete: ... - def parse_drs(self) -> Incomplete: ... - def parse_variable(self) -> Incomplete: ... - def parse_index(self) -> Incomplete: ... - -class BoxerDrsParser(DrtParser): - discourse_id: Incomplete - def __init__(self, discourse_id: Incomplete | None = None) -> None: ... - def get_all_symbols(self) -> Incomplete: ... - def attempt_adjuncts( - self, expression: Incomplete, context: Incomplete - ) -> Incomplete: ... - def handle(self, tok: Incomplete, context: Incomplete) -> Incomplete: ... - def nullableIntToken(self) -> Incomplete: ... - def get_next_token_variable( - self, description: Incomplete - ) -> Incomplete: ... - -class AbstractBoxerDrs: - def variables(self) -> Incomplete: ... - def variable_types(self) -> Incomplete: ... - def atoms(self) -> Incomplete: ... - def clean(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class BoxerDrs(AbstractBoxerDrs): - refs: Incomplete - conds: Incomplete - consequent: Incomplete - def __init__( - self, - refs: Incomplete, - conds: Incomplete, - consequent: Incomplete | None = None, - ) -> None: ... - def atoms(self) -> Incomplete: ... - def clean(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - -class BoxerNot(AbstractBoxerDrs): - drs: Incomplete - def __init__(self, drs: Incomplete) -> None: ... - def atoms(self) -> Incomplete: ... - def clean(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - -class BoxerIndexed(AbstractBoxerDrs): - discourse_id: Incomplete - sent_index: Incomplete - word_indices: Incomplete - def __init__( - self, - discourse_id: Incomplete, - sent_index: Incomplete, - word_indices: Incomplete, - ) -> None: ... - def atoms(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - -class BoxerPred(BoxerIndexed): - var: Incomplete - name: Incomplete - pos: Incomplete - sense: Incomplete - def __init__( - self, - discourse_id: Incomplete, - sent_index: Incomplete, - word_indices: Incomplete, - var: Incomplete, - name: Incomplete, - pos: Incomplete, - sense: Incomplete, - ) -> None: ... - def change_var(self, var: Incomplete) -> Incomplete: ... - def clean(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __iter__(self) -> Incomplete: ... - -class BoxerNamed(BoxerIndexed): - var: Incomplete - name: Incomplete - type: Incomplete - sense: Incomplete - def __init__( - self, - discourse_id: Incomplete, - sent_index: Incomplete, - word_indices: Incomplete, - var: Incomplete, - name: Incomplete, - type: Incomplete, - sense: Incomplete, - ) -> None: ... - def change_var(self, var: Incomplete) -> Incomplete: ... - def clean(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __iter__(self) -> Incomplete: ... - -class BoxerRel(BoxerIndexed): - var1: Incomplete - var2: Incomplete - rel: Incomplete - sense: Incomplete - def __init__( - self, - discourse_id: Incomplete, - sent_index: Incomplete, - word_indices: Incomplete, - var1: Incomplete, - var2: Incomplete, - rel: Incomplete, - sense: Incomplete, - ) -> None: ... - def clean(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __iter__(self) -> Incomplete: ... - -class BoxerProp(BoxerIndexed): - var: Incomplete - drs: Incomplete - def __init__( - self, - discourse_id: Incomplete, - sent_index: Incomplete, - word_indices: Incomplete, - var: Incomplete, - drs: Incomplete, - ) -> None: ... - def referenced_labels(self) -> Incomplete: ... - def atoms(self) -> Incomplete: ... - def clean(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __iter__(self) -> Incomplete: ... - -class BoxerEq(BoxerIndexed): - var1: Incomplete - var2: Incomplete - def __init__( - self, - discourse_id: Incomplete, - sent_index: Incomplete, - word_indices: Incomplete, - var1: Incomplete, - var2: Incomplete, - ) -> None: ... - def atoms(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __iter__(self) -> Incomplete: ... - -class BoxerCard(BoxerIndexed): - var: Incomplete - value: Incomplete - type: Incomplete - def __init__( - self, - discourse_id: Incomplete, - sent_index: Incomplete, - word_indices: Incomplete, - var: Incomplete, - value: Incomplete, - type: Incomplete, - ) -> None: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __iter__(self) -> Incomplete: ... - -class BoxerOr(BoxerIndexed): - drs1: Incomplete - drs2: Incomplete - def __init__( - self, - discourse_id: Incomplete, - sent_index: Incomplete, - word_indices: Incomplete, - drs1: Incomplete, - drs2: Incomplete, - ) -> None: ... - def atoms(self) -> Incomplete: ... - def clean(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __iter__(self) -> Incomplete: ... - -class BoxerWhq(BoxerIndexed): - ans_types: Incomplete - drs1: Incomplete - variable: Incomplete - drs2: Incomplete - def __init__( - self, - discourse_id: Incomplete, - sent_index: Incomplete, - word_indices: Incomplete, - ans_types: Incomplete, - drs1: Incomplete, - variable: Incomplete, - drs2: Incomplete, - ) -> None: ... - def atoms(self) -> Incomplete: ... - def clean(self) -> Incomplete: ... - def renumber_sentences(self, f: Incomplete) -> Incomplete: ... - def __iter__(self) -> Incomplete: ... - -class PassthroughBoxerDrsInterpreter: - def interpret(self, ex: Incomplete) -> Incomplete: ... - -class NltkDrtBoxerDrsInterpreter: - def __init__(self, occur_index: bool = False) -> None: ... - def interpret(self, ex: Incomplete) -> Incomplete: ... - -class UnparseableInputException(Exception): ... diff --git a/stubs/nltk/sem/chat80.pyi b/stubs/nltk/sem/chat80.pyi deleted file mode 100644 index 292d2f0..0000000 --- a/stubs/nltk/sem/chat80.pyi +++ /dev/null @@ -1,71 +0,0 @@ -from _typeshed import Incomplete - -borders: Incomplete -contains: Incomplete -city: Incomplete -country: Incomplete -circle_of_lat: Incomplete -circle_of_long: Incomplete -continent: Incomplete -region: Incomplete -ocean: Incomplete -sea: Incomplete -items: Incomplete -item_metadata: Incomplete -rels: Incomplete -not_unary: Incomplete - -class Concept: - prefLabel: Incomplete - arity: Incomplete - altLabels: Incomplete - closures: Incomplete - extension: Incomplete - def __init__( - self, - prefLabel: Incomplete, - arity: Incomplete, - altLabels: Incomplete = [], - closures: Incomplete = [], - extension: Incomplete = ..., - ) -> None: ... - def augment(self, data: Incomplete) -> Incomplete: ... - def close(self) -> None: ... - -def clause2concepts( - filename: Incomplete, - rel_name: Incomplete, - schema: Incomplete, - closures: Incomplete = [], -) -> Incomplete: ... -def cities2table( - filename: Incomplete, - rel_name: Incomplete, - dbname: Incomplete, - verbose: bool = False, - setup: bool = False, -) -> None: ... -def sql_query(dbname: Incomplete, query: Incomplete) -> Incomplete: ... -def unary_concept( - label: Incomplete, subj: Incomplete, records: Incomplete -) -> Incomplete: ... -def binary_concept( - label: Incomplete, - closures: Incomplete, - subj: Incomplete, - obj: Incomplete, - records: Incomplete, -) -> Incomplete: ... -def process_bundle(rels: Incomplete) -> Incomplete: ... -def make_valuation( - concepts: Incomplete, read: bool = False, lexicon: bool = False -) -> Incomplete: ... -def val_dump(rels: Incomplete, db: Incomplete) -> None: ... -def val_load(db: Incomplete) -> Incomplete: ... -def label_indivs( - valuation: Incomplete, lexicon: bool = False -) -> Incomplete: ... -def make_lex(symbols: Incomplete) -> Incomplete: ... -def concepts(items: Incomplete = ...) -> Incomplete: ... -def main() -> None: ... -def sql_demo() -> None: ... diff --git a/stubs/nltk/sem/cooper_storage.pyi b/stubs/nltk/sem/cooper_storage.pyi deleted file mode 100644 index dad3c47..0000000 --- a/stubs/nltk/sem/cooper_storage.pyi +++ /dev/null @@ -1,28 +0,0 @@ -from _typeshed import Incomplete - -from nltk.parse import load_parser as load_parser -from nltk.parse.featurechart import ( - InstantiateVarsChart as InstantiateVarsChart, -) -from nltk.sem.logic import ( - ApplicationExpression as ApplicationExpression, -) -from nltk.sem.logic import ( - LambdaExpression as LambdaExpression, -) -from nltk.sem.logic import ( - Variable as Variable, -) - -class CooperStore: - featstruct: Incomplete - readings: Incomplete - core: Incomplete - store: Incomplete - def __init__(self, featstruct: Incomplete) -> None: ... - def s_retrieve(self, trace: bool = False) -> None: ... - -def parse_with_bindops( - sentence: Incomplete, grammar: Incomplete | None = None, trace: int = 0 -) -> Incomplete: ... -def demo() -> None: ... diff --git a/stubs/nltk/sem/drt.pyi b/stubs/nltk/sem/drt.pyi deleted file mode 100644 index 4d5473e..0000000 --- a/stubs/nltk/sem/drt.pyi +++ /dev/null @@ -1,306 +0,0 @@ -from _typeshed import Incomplete - -from nltk.sem.logic import ( - APP as APP, -) -from nltk.sem.logic import ( - AbstractVariableExpression as AbstractVariableExpression, -) -from nltk.sem.logic import ( - AllExpression as AllExpression, -) -from nltk.sem.logic import ( - AndExpression as AndExpression, -) -from nltk.sem.logic import ( - ApplicationExpression as ApplicationExpression, -) -from nltk.sem.logic import ( - BinaryExpression as BinaryExpression, -) -from nltk.sem.logic import ( - BooleanExpression as BooleanExpression, -) -from nltk.sem.logic import ( - ConstantExpression as ConstantExpression, -) -from nltk.sem.logic import ( - EqualityExpression as EqualityExpression, -) -from nltk.sem.logic import ( - EventVariableExpression as EventVariableExpression, -) -from nltk.sem.logic import ( - ExistsExpression as ExistsExpression, -) -from nltk.sem.logic import ( - Expression as Expression, -) -from nltk.sem.logic import ( - FunctionVariableExpression as FunctionVariableExpression, -) -from nltk.sem.logic import ( - ImpExpression as ImpExpression, -) -from nltk.sem.logic import ( - IndividualVariableExpression as IndividualVariableExpression, -) -from nltk.sem.logic import ( - LambdaExpression as LambdaExpression, -) -from nltk.sem.logic import ( - LogicParser as LogicParser, -) -from nltk.sem.logic import ( - NegatedExpression as NegatedExpression, -) -from nltk.sem.logic import ( - OrExpression as OrExpression, -) -from nltk.sem.logic import ( - Tokens as tokens, -) -from nltk.sem.logic import ( - Variable as Variable, -) -from nltk.sem.logic import ( - is_eventvar as is_eventvar, -) -from nltk.sem.logic import ( - is_funcvar as is_funcvar, -) -from nltk.sem.logic import ( - is_indvar as is_indvar, -) -from nltk.sem.logic import ( - unique_variable as unique_variable, -) -from nltk.util import in_idle as in_idle - -class DrtTokens(tokens): - DRS: str - DRS_CONC: str - PRONOUN: str - OPEN_BRACKET: str - CLOSE_BRACKET: str - COLON: str - PUNCT: Incomplete - SYMBOLS: Incomplete - TOKENS: Incomplete - -class DrtParser(LogicParser): - operator_precedence: Incomplete - def __init__(self) -> None: ... - def get_all_symbols(self) -> Incomplete: ... - def isvariable(self, tok: Incomplete) -> Incomplete: ... - def handle(self, tok: Incomplete, context: Incomplete) -> Incomplete: ... - def make_NegatedExpression(self, expression: Incomplete) -> Incomplete: ... - def handle_DRS( - self, tok: Incomplete, context: Incomplete - ) -> Incomplete: ... - def handle_refs(self) -> Incomplete: ... - def handle_conds(self, context: Incomplete) -> Incomplete: ... - def handle_prop( - self, tok: Incomplete, context: Incomplete - ) -> Incomplete: ... - def make_EqualityExpression( - self, first: Incomplete, second: Incomplete - ) -> Incomplete: ... - def get_BooleanExpression_factory(self, tok: Incomplete) -> Incomplete: ... - def make_BooleanExpression( - self, factory: Incomplete, first: Incomplete, second: Incomplete - ) -> Incomplete: ... - def make_ApplicationExpression( - self, function: Incomplete, argument: Incomplete - ) -> Incomplete: ... - def make_VariableExpression(self, name: Incomplete) -> Incomplete: ... - def make_LambdaExpression( - self, variables: Incomplete, term: Incomplete - ) -> Incomplete: ... - -class DrtExpression: - @classmethod - def fromstring(cls: Incomplete, s: Incomplete) -> Incomplete: ... - def applyto(self, other: Incomplete) -> Incomplete: ... - def __neg__(self) -> Incomplete: ... - def __and__(self, other: Incomplete) -> Incomplete: ... - def __or__(self, other: Incomplete) -> Incomplete: ... - def __gt__(self, other: Incomplete) -> Incomplete: ... - def equiv( - self, other: Incomplete, prover: Incomplete | None = None - ) -> Incomplete: ... - @property - def type(self) -> None: ... - def typecheck(self, signature: Incomplete | None = None) -> None: ... - def __add__(self, other: Incomplete) -> Incomplete: ... - def get_refs(self, recursive: bool = False) -> None: ... - def is_pronoun_function(self) -> Incomplete: ... - def make_EqualityExpression( - self, first: Incomplete, second: Incomplete - ) -> Incomplete: ... - def make_VariableExpression(self, variable: Incomplete) -> Incomplete: ... - def resolve_anaphora(self) -> Incomplete: ... - def eliminate_equality(self) -> Incomplete: ... - def pretty_format(self) -> Incomplete: ... - def pretty_print(self) -> None: ... - def draw(self) -> None: ... - -class DRS(DrtExpression, Expression): - refs: Incomplete - conds: Incomplete - consequent: Incomplete - def __init__( - self, - refs: Incomplete, - conds: Incomplete, - consequent: Incomplete | None = None, - ) -> None: ... - def replace( - self, - variable: Incomplete, - expression: Incomplete, - replace_bound: bool = False, - alpha_convert: bool = True, - ) -> Incomplete: ... - def free(self) -> Incomplete: ... - def get_refs(self, recursive: bool = False) -> Incomplete: ... - def visit( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - def visit_structured( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - def eliminate_equality(self) -> Incomplete: ... - def fol(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - -def DrtVariableExpression(variable: Incomplete) -> Incomplete: ... - -class DrtAbstractVariableExpression(DrtExpression, AbstractVariableExpression): - def fol(self) -> Incomplete: ... - def get_refs(self, recursive: bool = False) -> Incomplete: ... - def eliminate_equality(self) -> Incomplete: ... - -class DrtIndividualVariableExpression( - DrtAbstractVariableExpression, IndividualVariableExpression -): ... -class DrtFunctionVariableExpression( - DrtAbstractVariableExpression, FunctionVariableExpression -): ... -class DrtEventVariableExpression( - DrtIndividualVariableExpression, EventVariableExpression -): ... -class DrtConstantExpression( - DrtAbstractVariableExpression, ConstantExpression -): ... - -class DrtProposition(DrtExpression, Expression): - variable: Incomplete - drs: Incomplete - def __init__(self, variable: Incomplete, drs: Incomplete) -> None: ... - def replace( - self, - variable: Incomplete, - expression: Incomplete, - replace_bound: bool = False, - alpha_convert: bool = True, - ) -> Incomplete: ... - def eliminate_equality(self) -> Incomplete: ... - def get_refs(self, recursive: bool = False) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - def fol(self) -> Incomplete: ... - def visit( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - def visit_structured( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - -class DrtNegatedExpression(DrtExpression, NegatedExpression): - def fol(self) -> Incomplete: ... - def get_refs(self, recursive: bool = False) -> Incomplete: ... - -class DrtLambdaExpression(DrtExpression, LambdaExpression): - def alpha_convert(self, newvar: Incomplete) -> Incomplete: ... - def fol(self) -> Incomplete: ... - def get_refs(self, recursive: bool = False) -> Incomplete: ... - -class DrtBinaryExpression(DrtExpression, BinaryExpression): - def get_refs(self, recursive: bool = False) -> Incomplete: ... - -class DrtBooleanExpression(DrtBinaryExpression, BooleanExpression): ... - -class DrtOrExpression(DrtBooleanExpression, OrExpression): - def fol(self) -> Incomplete: ... - -class DrtEqualityExpression(DrtBinaryExpression, EqualityExpression): - def fol(self) -> Incomplete: ... - -class DrtConcatenation(DrtBooleanExpression): - consequent: Incomplete - def __init__( - self, - first: Incomplete, - second: Incomplete, - consequent: Incomplete | None = None, - ) -> None: ... - def replace( - self, - variable: Incomplete, - expression: Incomplete, - replace_bound: bool = False, - alpha_convert: bool = True, - ) -> Incomplete: ... - def eliminate_equality(self) -> Incomplete: ... - def simplify(self) -> Incomplete: ... - def get_refs(self, recursive: bool = False) -> Incomplete: ... - def getOp(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - def fol(self) -> Incomplete: ... - def visit( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - -class DrtApplicationExpression(DrtExpression, ApplicationExpression): - def fol(self) -> Incomplete: ... - def get_refs(self, recursive: bool = False) -> Incomplete: ... - -class PossibleAntecedents(list, DrtExpression, Expression): - def free(self) -> Incomplete: ... - def replace( - self, - variable: Incomplete, - expression: Incomplete, - replace_bound: bool = False, - alpha_convert: bool = True, - ) -> Incomplete: ... - -class AnaphoraResolutionException(Exception): ... - -def resolve_anaphora( - expression: Incomplete, trail: Incomplete = [] -) -> Incomplete: ... - -class DrsDrawer: - BUFFER: int - TOPSPACE: int - OUTERSPACE: int - canvas: Incomplete - drs: Incomplete - master: Incomplete - def __init__( - self, - dr: Incomplete, - size_canvas: bool = True, - canvas: Incomplete | None = None, - ) -> None: ... - def draw(self, x: Incomplete = ..., y: Incomplete = ...) -> Incomplete: ... - -def demo() -> None: ... -def test_draw() -> None: ... diff --git a/stubs/nltk/sem/drt_glue_demo.pyi b/stubs/nltk/sem/drt_glue_demo.pyi deleted file mode 100644 index ceec1b8..0000000 --- a/stubs/nltk/sem/drt_glue_demo.pyi +++ /dev/null @@ -1,31 +0,0 @@ -from _typeshed import Incomplete - -from nltk.draw.util import CanvasFrame as CanvasFrame -from nltk.draw.util import ShowText as ShowText -from nltk.parse import MaltParser as MaltParser -from nltk.sem.drt import DrsDrawer as DrsDrawer -from nltk.sem.drt import DrtVariableExpression as DrtVariableExpression -from nltk.sem.glue import DrtGlue as DrtGlue -from nltk.sem.logic import Variable as Variable -from nltk.tag import RegexpTagger as RegexpTagger -from nltk.util import in_idle as in_idle - -class DrtGlueDemo: - def __init__(self, examples: Incomplete) -> None: ... - def destroy(self, *e: Incomplete) -> None: ... - def prev(self, *e: Incomplete) -> None: ... - def next(self, *e: Incomplete) -> None: ... - def about(self, *e: Incomplete) -> None: ... - def postscript(self, *e: Incomplete) -> None: ... - def mainloop(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def resize(self, size: Incomplete | None = None) -> None: ... - -class DrsWidget: - bbox: Incomplete - def __init__( - self, canvas: Incomplete, drs: Incomplete, **attribs: Incomplete - ) -> None: ... - def draw(self) -> None: ... - def clear(self) -> None: ... - -def demo() -> None: ... diff --git a/stubs/nltk/sem/evaluate.pyi b/stubs/nltk/sem/evaluate.pyi deleted file mode 100644 index 8a7872b..0000000 --- a/stubs/nltk/sem/evaluate.pyi +++ /dev/null @@ -1,119 +0,0 @@ -from _typeshed import Incomplete - -from nltk.decorators import decorator as decorator -from nltk.sem.logic import ( - AbstractVariableExpression as AbstractVariableExpression, -) -from nltk.sem.logic import ( - AllExpression as AllExpression, -) -from nltk.sem.logic import ( - AndExpression as AndExpression, -) -from nltk.sem.logic import ( - ApplicationExpression as ApplicationExpression, -) -from nltk.sem.logic import ( - EqualityExpression as EqualityExpression, -) -from nltk.sem.logic import ( - ExistsExpression as ExistsExpression, -) -from nltk.sem.logic import ( - Expression as Expression, -) -from nltk.sem.logic import ( - IffExpression as IffExpression, -) -from nltk.sem.logic import ( - ImpExpression as ImpExpression, -) -from nltk.sem.logic import ( - IndividualVariableExpression as IndividualVariableExpression, -) -from nltk.sem.logic import ( - IotaExpression as IotaExpression, -) -from nltk.sem.logic import ( - LambdaExpression as LambdaExpression, -) -from nltk.sem.logic import ( - NegatedExpression as NegatedExpression, -) -from nltk.sem.logic import ( - OrExpression as OrExpression, -) -from nltk.sem.logic import ( - Variable as Variable, -) -from nltk.sem.logic import ( - is_indvar as is_indvar, -) - -class Error(Exception): ... -class Undefined(Error): ... - -def trace( - f: Incomplete, *args: Incomplete, **kw: Incomplete -) -> Incomplete: ... -def is_rel(s: Incomplete) -> Incomplete: ... -def set2rel(s: Incomplete) -> Incomplete: ... -def arity(rel: Incomplete) -> Incomplete: ... - -class Valuation(dict): - def __init__(self, xs: Incomplete) -> None: ... - def __getitem__(self, key: Incomplete) -> Incomplete: ... - @property - def domain(self) -> Incomplete: ... - @property - def symbols(self) -> Incomplete: ... - @classmethod - def fromstring(cls: Incomplete, s: Incomplete) -> Incomplete: ... - -def read_valuation( - s: Incomplete, encoding: Incomplete | None = None -) -> Incomplete: ... - -class Assignment(dict): - domain: Incomplete - variant: Incomplete - def __init__( - self, domain: Incomplete, assign: Incomplete | None = None - ) -> None: ... - def __getitem__(self, key: Incomplete) -> Incomplete: ... - def copy(self) -> Incomplete: ... - def purge(self, var: Incomplete | None = None) -> None: ... - def add(self, var: Incomplete, val: Incomplete) -> Incomplete: ... - -class Model: - domain: Incomplete - valuation: Incomplete - def __init__(self, domain: Incomplete, valuation: Incomplete) -> None: ... - def evaluate( - self, expr: Incomplete, g: Incomplete, trace: Incomplete | None = None - ) -> Incomplete: ... - def satisfy( - self, - parsed: Incomplete, - g: Incomplete, - trace: Incomplete | None = None, - ) -> Incomplete: ... - def i( - self, parsed: Incomplete, g: Incomplete, trace: bool = False - ) -> Incomplete: ... - def satisfiers( - self, - parsed: Incomplete, - varex: Incomplete, - g: Incomplete, - trace: Incomplete | None = None, - nesting: int = 0, - ) -> Incomplete: ... - -mult: int - -def propdemo(trace: Incomplete | None = None) -> None: ... -def folmodel(quiet: bool = False, trace: Incomplete | None = None) -> None: ... -def foldemo(trace: Incomplete | None = None) -> None: ... -def satdemo(trace: Incomplete | None = None) -> None: ... -def demo(num: int = 0, trace: Incomplete | None = None) -> None: ... diff --git a/stubs/nltk/sem/glue.pyi b/stubs/nltk/sem/glue.pyi deleted file mode 100644 index 0e8c1a7..0000000 --- a/stubs/nltk/sem/glue.pyi +++ /dev/null @@ -1,146 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import Counter as Counter -from nltk.sem import drt as drt -from nltk.sem import linearlogic as linearlogic -from nltk.sem.logic import ( - AbstractVariableExpression as AbstractVariableExpression, -) -from nltk.sem.logic import Expression as Expression -from nltk.sem.logic import LambdaExpression as LambdaExpression -from nltk.sem.logic import Variable as Variable -from nltk.sem.logic import VariableExpression as VariableExpression -from nltk.tag import BigramTagger as BigramTagger -from nltk.tag import RegexpTagger as RegexpTagger -from nltk.tag import TrigramTagger as TrigramTagger -from nltk.tag import UnigramTagger as UnigramTagger - -SPEC_SEMTYPES: Incomplete -OPTIONAL_RELATIONSHIPS: Incomplete - -class GlueFormula: - meaning: Incomplete - glue: Incomplete - indices: Incomplete - def __init__( - self, - meaning: Incomplete, - glue: Incomplete, - indices: Incomplete | None = None, - ) -> None: ... - def applyto(self, arg: Incomplete) -> Incomplete: ... - def make_VariableExpression(self, name: Incomplete) -> Incomplete: ... - def make_LambdaExpression( - self: Incomplete, variable: Incomplete, term: Incomplete - ) -> Incomplete: ... - def lambda_abstract(self, other: Incomplete) -> Incomplete: ... - def compile(self, counter: Incomplete | None = None) -> Incomplete: ... - def simplify(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - -class GlueDict(dict): - filename: Incomplete - file_encoding: Incomplete - def __init__( - self, filename: Incomplete, encoding: Incomplete | None = None - ) -> None: ... - def read_file(self, empty_first: bool = True) -> None: ... - def to_glueformula_list( - self, - depgraph: Incomplete, - node: Incomplete | None = None, - counter: Incomplete | None = None, - verbose: bool = False, - ) -> Incomplete: ... - def lookup( - self, node: Incomplete, depgraph: Incomplete, counter: Incomplete - ) -> Incomplete: ... - def add_missing_dependencies( - self, node: Incomplete, depgraph: Incomplete - ) -> None: ... - def get_semtypes(self, node: Incomplete) -> Incomplete: ... - def get_glueformulas_from_semtype_entry( - self, - lookup: Incomplete, - word: Incomplete, - node: Incomplete, - depgraph: Incomplete, - counter: Incomplete, - ) -> Incomplete: ... - def get_meaning_formula( - self, generic: Incomplete, word: Incomplete - ) -> Incomplete: ... - def initialize_labels( - self, - expr: Incomplete, - node: Incomplete, - depgraph: Incomplete, - unique_index: Incomplete, - ) -> Incomplete: ... - def find_label_name( - self, - name: Incomplete, - node: Incomplete, - depgraph: Incomplete, - unique_index: Incomplete, - ) -> Incomplete: ... - def get_label(self, node: Incomplete) -> Incomplete: ... - def lookup_unique( - self, rel: Incomplete, node: Incomplete, depgraph: Incomplete - ) -> Incomplete: ... - def get_GlueFormula_factory(self) -> Incomplete: ... - -class Glue: - verbose: Incomplete - remove_duplicates: Incomplete - depparser: Incomplete - prover: Incomplete - semtype_file: Incomplete - def __init__( - self, - semtype_file: Incomplete | None = None, - remove_duplicates: bool = False, - depparser: Incomplete | None = None, - verbose: bool = False, - ) -> None: ... - def train_depparser(self, depgraphs: Incomplete | None = None) -> None: ... - def parse_to_meaning(self, sentence: Incomplete) -> Incomplete: ... - def get_readings(self, agenda: Incomplete) -> Incomplete: ... - def parse_to_compiled(self, sentence: Incomplete) -> Incomplete: ... - def dep_parse(self, sentence: Incomplete) -> Incomplete: ... - def depgraph_to_glue(self, depgraph: Incomplete) -> Incomplete: ... - def get_glue_dict(self) -> Incomplete: ... - def gfl_to_compiled(self, gfl: Incomplete) -> Incomplete: ... - def get_pos_tagger(self) -> Incomplete: ... - -class DrtGlueFormula(GlueFormula): - meaning: Incomplete - glue: Incomplete - indices: Incomplete - def __init__( - self, - meaning: Incomplete, - glue: Incomplete, - indices: Incomplete | None = None, - ) -> None: ... - def make_VariableExpression(self, name: Incomplete) -> Incomplete: ... - def make_LambdaExpression( - self, variable: Incomplete, term: Incomplete - ) -> Incomplete: ... - -class DrtGlueDict(GlueDict): - def get_GlueFormula_factory(self) -> Incomplete: ... - -class DrtGlue(Glue): - def __init__( - self, - semtype_file: Incomplete | None = None, - remove_duplicates: bool = False, - depparser: Incomplete | None = None, - verbose: bool = False, - ) -> None: ... - def get_glue_dict(self) -> Incomplete: ... - -def demo(show_example: int = -1) -> None: ... diff --git a/stubs/nltk/sem/hole.pyi b/stubs/nltk/sem/hole.pyi deleted file mode 100644 index 271a770..0000000 --- a/stubs/nltk/sem/hole.pyi +++ /dev/null @@ -1,71 +0,0 @@ -from _typeshed import Incomplete - -from nltk.parse import load_parser as load_parser -from nltk.sem.logic import ( - AllExpression as AllExpression, -) -from nltk.sem.logic import ( - AndExpression as AndExpression, -) -from nltk.sem.logic import ( - ApplicationExpression as ApplicationExpression, -) -from nltk.sem.logic import ( - ExistsExpression as ExistsExpression, -) -from nltk.sem.logic import ( - IffExpression as IffExpression, -) -from nltk.sem.logic import ( - ImpExpression as ImpExpression, -) -from nltk.sem.logic import ( - LambdaExpression as LambdaExpression, -) -from nltk.sem.logic import ( - NegatedExpression as NegatedExpression, -) -from nltk.sem.logic import ( - OrExpression as OrExpression, -) -from nltk.sem.skolemize import skolemize as skolemize - -class Constants: - ALL: str - EXISTS: str - NOT: str - AND: str - OR: str - IMP: str - IFF: str - PRED: str - LEQ: str - HOLE: str - LABEL: str - MAP: Incomplete - -class HoleSemantics: - holes: Incomplete - labels: Incomplete - fragments: Incomplete - constraints: Incomplete - top_most_labels: Incomplete - top_hole: Incomplete - def __init__(self, usr: Incomplete) -> None: ... - def is_node(self, x: Incomplete) -> Incomplete: ... - def pluggings(self) -> Incomplete: ... - def formula_tree(self, plugging: Incomplete) -> Incomplete: ... - -class Constraint: - lhs: Incomplete - rhs: Incomplete - def __init__(self, lhs: Incomplete, rhs: Incomplete) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -def hole_readings( - sentence: Incomplete, - grammar_filename: Incomplete | None = None, - verbose: bool = False, -) -> Incomplete: ... diff --git a/stubs/nltk/sem/lfg.pyi b/stubs/nltk/sem/lfg.pyi deleted file mode 100644 index 2bb8171..0000000 --- a/stubs/nltk/sem/lfg.pyi +++ /dev/null @@ -1,16 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import Counter as Counter - -class FStructure(dict): - def safeappend(self, key: Incomplete, item: Incomplete) -> None: ... - def __setitem__(self, key: Incomplete, value: Incomplete) -> None: ... - def __getitem__(self, key: Incomplete) -> Incomplete: ... - def __contains__(self, key: Incomplete) -> bool: ... - def to_glueformula_list(self, glue_dict: Incomplete) -> Incomplete: ... - def to_depgraph(self, rel: Incomplete | None = None) -> Incomplete: ... - @staticmethod - def read_depgraph(depgraph: Incomplete) -> Incomplete: ... - def pretty_format(self, indent: int = 3) -> Incomplete: ... - -def demo_read_depgraph() -> None: ... diff --git a/stubs/nltk/sem/linearlogic.pyi b/stubs/nltk/sem/linearlogic.pyi deleted file mode 100644 index fce22b8..0000000 --- a/stubs/nltk/sem/linearlogic.pyi +++ /dev/null @@ -1,115 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import Counter as Counter -from nltk.sem.logic import APP as APP -from nltk.sem.logic import LogicParser as LogicParser - -class Tokens: - OPEN: str - CLOSE: str - IMP: str - PUNCT: Incomplete - TOKENS: Incomplete - -class LinearLogicParser(LogicParser): - operator_precedence: Incomplete - def __init__(self) -> None: ... - def get_all_symbols(self) -> Incomplete: ... - def handle(self, tok: Incomplete, context: Incomplete) -> Incomplete: ... - def get_BooleanExpression_factory(self, tok: Incomplete) -> Incomplete: ... - def make_BooleanExpression( - self, factory: Incomplete, first: Incomplete, second: Incomplete - ) -> Incomplete: ... - def attempt_ApplicationExpression( - self, expression: Incomplete, context: Incomplete - ) -> Incomplete: ... - def make_VariableExpression(self, name: Incomplete) -> Incomplete: ... - -class Expression: - @classmethod - def fromstring(cls: Incomplete, s: Incomplete) -> Incomplete: ... - def applyto( - self, other: Incomplete, other_indices: Incomplete | None = None - ) -> Incomplete: ... - def __call__(self, other: Incomplete) -> Incomplete: ... - -class AtomicExpression(Expression): - name: Incomplete - dependencies: Incomplete - def __init__( - self, name: Incomplete, dependencies: Incomplete | None = None - ) -> None: ... - def simplify(self, bindings: Incomplete | None = None) -> Incomplete: ... - def compile_pos( - self, index_counter: Incomplete, glueFormulaFactory: Incomplete - ) -> Incomplete: ... - def compile_neg( - self, index_counter: Incomplete, glueFormulaFactory: Incomplete - ) -> Incomplete: ... - def initialize_labels(self, fstruct: Incomplete) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class ConstantExpression(AtomicExpression): - def unify(self, other: Incomplete, bindings: Incomplete) -> Incomplete: ... - -class VariableExpression(AtomicExpression): - def unify(self, other: Incomplete, bindings: Incomplete) -> Incomplete: ... - -class ImpExpression(Expression): - antecedent: Incomplete - consequent: Incomplete - def __init__( - self, antecedent: Incomplete, consequent: Incomplete - ) -> None: ... - def simplify(self, bindings: Incomplete | None = None) -> Incomplete: ... - def unify(self, other: Incomplete, bindings: Incomplete) -> Incomplete: ... - def compile_pos( - self, index_counter: Incomplete, glueFormulaFactory: Incomplete - ) -> Incomplete: ... - def compile_neg( - self, index_counter: Incomplete, glueFormulaFactory: Incomplete - ) -> Incomplete: ... - def initialize_labels(self, fstruct: Incomplete) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class ApplicationExpression(Expression): - function: Incomplete - argument: Incomplete - bindings: Incomplete - def __init__( - self, - function: Incomplete, - argument: Incomplete, - argument_indices: Incomplete | None = None, - ) -> None: ... - def simplify(self, bindings: Incomplete | None = None) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class BindingDict: - d: Incomplete - def __init__(self, bindings: Incomplete | None = None) -> None: ... - def __setitem__( - self, variable: Incomplete, binding: Incomplete - ) -> None: ... - def __getitem__(self, variable: Incomplete) -> Incomplete: ... - def __contains__(self, item: Incomplete) -> bool: ... - def __add__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - -class VariableBindingException(Exception): ... - -class UnificationException(Exception): - def __init__( - self, a: Incomplete, b: Incomplete, bindings: Incomplete - ) -> None: ... - -class LinearLogicApplicationException(Exception): ... - -def demo() -> None: ... diff --git a/stubs/nltk/sem/logic.pyi b/stubs/nltk/sem/logic.pyi deleted file mode 100644 index dd2444d..0000000 --- a/stubs/nltk/sem/logic.pyi +++ /dev/null @@ -1,439 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import Counter as Counter -from nltk.util import Trie as Trie - -APP: str - -class Tokens: - LAMBDA: str - LAMBDA_LIST: Incomplete - EXISTS: str - EXISTS_LIST: Incomplete - ALL: str - ALL_LIST: Incomplete - IOTA: str - IOTA_LIST: Incomplete - DOT: str - OPEN: str - CLOSE: str - COMMA: str - NOT: str - NOT_LIST: Incomplete - AND: str - AND_LIST: Incomplete - OR: str - OR_LIST: Incomplete - IMP: str - IMP_LIST: Incomplete - IFF: str - IFF_LIST: Incomplete - EQ: str - EQ_LIST: Incomplete - NEQ: str - NEQ_LIST: Incomplete - BINOPS: Incomplete - QUANTS: Incomplete - PUNCT: Incomplete - TOKENS: Incomplete - SYMBOLS: Incomplete - -def boolean_ops() -> None: ... -def equality_preds() -> None: ... -def binding_ops() -> None: ... - -class LogicParser: - type_check: Incomplete - quote_chars: Incomplete - operator_precedence: Incomplete - right_associated_operations: Incomplete - def __init__(self, type_check: bool = False) -> None: ... - def parse( - self, data: Incomplete, signature: Incomplete | None = None - ) -> Incomplete: ... - def process(self, data: Incomplete) -> Incomplete: ... - def process_quoted_token( - self, data_idx: Incomplete, data: Incomplete - ) -> Incomplete: ... - def get_all_symbols(self) -> Incomplete: ... - def inRange(self, location: Incomplete) -> Incomplete: ... - def token(self, location: Incomplete | None = None) -> Incomplete: ... - def isvariable(self, tok: Incomplete) -> Incomplete: ... - def process_next_expression(self, context: Incomplete) -> Incomplete: ... - def handle(self, tok: Incomplete, context: Incomplete) -> Incomplete: ... - def attempt_adjuncts( - self, expression: Incomplete, context: Incomplete - ) -> Incomplete: ... - def handle_negation( - self, tok: Incomplete, context: Incomplete - ) -> Incomplete: ... - def make_NegatedExpression(self, expression: Incomplete) -> Incomplete: ... - def handle_variable( - self, tok: Incomplete, context: Incomplete - ) -> Incomplete: ... - def get_next_token_variable( - self, description: Incomplete - ) -> Incomplete: ... - def handle_lambda( - self, tok: Incomplete, context: Incomplete - ) -> Incomplete: ... - def handle_quant( - self, tok: Incomplete, context: Incomplete - ) -> Incomplete: ... - def get_QuantifiedExpression_factory( - self, tok: Incomplete - ) -> Incomplete: ... - def make_QuanifiedExpression( - self, factory: Incomplete, variable: Incomplete, term: Incomplete - ) -> Incomplete: ... - def handle_open( - self, tok: Incomplete, context: Incomplete - ) -> Incomplete: ... - def attempt_EqualityExpression( - self, expression: Incomplete, context: Incomplete - ) -> Incomplete: ... - def make_EqualityExpression( - self, first: Incomplete, second: Incomplete - ) -> Incomplete: ... - def attempt_BooleanExpression( - self, expression: Incomplete, context: Incomplete - ) -> Incomplete: ... - def get_BooleanExpression_factory(self, tok: Incomplete) -> Incomplete: ... - def make_BooleanExpression( - self, factory: Incomplete, first: Incomplete, second: Incomplete - ) -> Incomplete: ... - def attempt_ApplicationExpression( - self, expression: Incomplete, context: Incomplete - ) -> Incomplete: ... - def make_ApplicationExpression( - self, function: Incomplete, argument: Incomplete - ) -> Incomplete: ... - def make_VariableExpression(self, name: Incomplete) -> Incomplete: ... - def make_LambdaExpression( - self, variable: Incomplete, term: Incomplete - ) -> Incomplete: ... - def has_priority( - self, operation: Incomplete, context: Incomplete - ) -> Incomplete: ... - def assertNextToken(self, expected: Incomplete) -> None: ... - def assertToken(self, tok: Incomplete, expected: Incomplete) -> None: ... - -def read_logic( - s: Incomplete, - logic_parser: Incomplete | None = None, - encoding: Incomplete | None = None, -) -> Incomplete: ... - -class Variable: - name: Incomplete - def __init__(self, name: Incomplete) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def substitute_bindings(self, bindings: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -def unique_variable( - pattern: Incomplete | None = None, ignore: Incomplete | None = None -) -> Incomplete: ... -def skolem_function(univ_scope: Incomplete | None = None) -> Incomplete: ... - -class Type: - def __hash__(self) -> Incomplete: ... - @classmethod - def fromstring(cls: Incomplete, s: Incomplete) -> Incomplete: ... - -class ComplexType(Type): - first: Incomplete - second: Incomplete - def __init__(self, first: Incomplete, second: Incomplete) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - def matches(self, other: Incomplete) -> Incomplete: ... - def resolve(self, other: Incomplete) -> Incomplete: ... - def str(self) -> Incomplete: ... - -class BasicType(Type): - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - def matches(self, other: Incomplete) -> Incomplete: ... - def resolve(self, other: Incomplete) -> Incomplete: ... - -class EntityType(BasicType): - def str(self) -> Incomplete: ... - -class TruthValueType(BasicType): - def str(self) -> Incomplete: ... - -class EventType(BasicType): - def str(self) -> Incomplete: ... - -class AnyType(BasicType, ComplexType): - def __init__(self) -> None: ... - @property - def first(self) -> Incomplete: ... - @property - def second(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - def matches(self, other: Incomplete) -> Incomplete: ... - def resolve(self, other: Incomplete) -> Incomplete: ... - def str(self) -> Incomplete: ... - -TRUTH_TYPE: Incomplete -ENTITY_TYPE: Incomplete -EVENT_TYPE: Incomplete -ANY_TYPE: Incomplete - -def read_type(type_string: Incomplete) -> Incomplete: ... - -class TypeException(Exception): - def __init__(self, msg: Incomplete) -> None: ... - -class InconsistentTypeHierarchyException(TypeException): - def __init__( - self, variable: Incomplete, expression: Incomplete | None = None - ) -> None: ... - -class TypeResolutionException(TypeException): - def __init__( - self, expression: Incomplete, other_type: Incomplete - ) -> None: ... - -class IllegalTypeException(TypeException): - def __init__( - self, - expression: Incomplete, - other_type: Incomplete, - allowed_type: Incomplete, - ) -> None: ... - -def typecheck( - expressions: Incomplete, signature: Incomplete | None = None -) -> Incomplete: ... - -class SubstituteBindingsI: - def substitute_bindings(self, bindings: Incomplete) -> None: ... - def variables(self) -> None: ... - -class Expression(SubstituteBindingsI): - @classmethod - def fromstring( - cls: Incomplete, - s: Incomplete, - type_check: bool = False, - signature: Incomplete | None = None, - ) -> Incomplete: ... - def __call__( - self, other: Incomplete, *additional: Incomplete - ) -> Incomplete: ... - def applyto(self, other: Incomplete) -> Incomplete: ... - def __neg__(self) -> Incomplete: ... - def negate(self) -> Incomplete: ... - def __and__(self, other: Incomplete) -> Incomplete: ... - def __or__(self, other: Incomplete) -> Incomplete: ... - def __gt__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def equiv( - self, other: Incomplete, prover: Incomplete | None = None - ) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - def substitute_bindings(self, bindings: Incomplete) -> Incomplete: ... - def typecheck(self, signature: Incomplete | None = None) -> Incomplete: ... - def findtype(self, variable: Incomplete) -> None: ... - def replace( - self, - variable: Incomplete, - expression: Incomplete, - replace_bound: bool = False, - alpha_convert: bool = True, - ) -> Incomplete: ... - def normalize(self, newvars: Incomplete | None = None) -> Incomplete: ... - def visit(self, function: Incomplete, combinator: Incomplete) -> None: ... - def visit_structured( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - def variables(self) -> Incomplete: ... - def free(self) -> Incomplete: ... - def constants(self) -> Incomplete: ... - def predicates(self) -> Incomplete: ... - def simplify(self) -> Incomplete: ... - def make_VariableExpression(self, variable: Incomplete) -> Incomplete: ... - -class ApplicationExpression(Expression): - function: Incomplete - argument: Incomplete - def __init__(self, function: Incomplete, argument: Incomplete) -> None: ... - def simplify(self) -> Incomplete: ... - @property - def type(self) -> Incomplete: ... - def findtype(self, variable: Incomplete) -> Incomplete: ... - def constants(self) -> Incomplete: ... - def predicates(self) -> Incomplete: ... - def visit( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - def uncurry(self) -> Incomplete: ... - @property - def pred(self) -> Incomplete: ... - @property - def args(self) -> Incomplete: ... - def is_atom(self) -> Incomplete: ... - -class AbstractVariableExpression(Expression): - variable: Incomplete - def __init__(self, variable: Incomplete) -> None: ... - def simplify(self) -> Incomplete: ... - def replace( - self, - variable: Incomplete, - expression: Incomplete, - replace_bound: bool = False, - alpha_convert: bool = True, - ) -> Incomplete: ... - def findtype(self, variable: Incomplete) -> Incomplete: ... - def predicates(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - -class IndividualVariableExpression(AbstractVariableExpression): - type: Incomplete - def free(self) -> Incomplete: ... - def constants(self) -> Incomplete: ... - -class FunctionVariableExpression(AbstractVariableExpression): - type = ANY_TYPE - def free(self) -> Incomplete: ... - def constants(self) -> Incomplete: ... - -class EventVariableExpression(IndividualVariableExpression): - type = EVENT_TYPE - -class ConstantExpression(AbstractVariableExpression): - type = ENTITY_TYPE - def free(self) -> Incomplete: ... - def constants(self) -> Incomplete: ... - -def VariableExpression(variable: Incomplete) -> Incomplete: ... - -class VariableBinderExpression(Expression): - variable: Incomplete - term: Incomplete - def __init__(self, variable: Incomplete, term: Incomplete) -> None: ... - def replace( - self, - variable: Incomplete, - expression: Incomplete, - replace_bound: bool = False, - alpha_convert: bool = True, - ) -> Incomplete: ... - def alpha_convert(self, newvar: Incomplete) -> Incomplete: ... - def free(self) -> Incomplete: ... - def findtype(self, variable: Incomplete) -> Incomplete: ... - def visit( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - def visit_structured( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - -class LambdaExpression(VariableBinderExpression): - @property - def type(self) -> Incomplete: ... - -class QuantifiedExpression(VariableBinderExpression): - @property - def type(self) -> Incomplete: ... - -class ExistsExpression(QuantifiedExpression): - def getQuantifier(self) -> Incomplete: ... - -class AllExpression(QuantifiedExpression): - def getQuantifier(self) -> Incomplete: ... - -class IotaExpression(QuantifiedExpression): - def getQuantifier(self) -> Incomplete: ... - -class NegatedExpression(Expression): - term: Incomplete - def __init__(self, term: Incomplete) -> None: ... - @property - def type(self) -> Incomplete: ... - def findtype(self, variable: Incomplete) -> Incomplete: ... - def visit( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - def negate(self) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - -class BinaryExpression(Expression): - first: Incomplete - second: Incomplete - def __init__(self, first: Incomplete, second: Incomplete) -> None: ... - @property - def type(self) -> Incomplete: ... - def findtype(self, variable: Incomplete) -> Incomplete: ... - def visit( - self, function: Incomplete, combinator: Incomplete - ) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - __hash__: Incomplete - -class BooleanExpression(BinaryExpression): ... - -class AndExpression(BooleanExpression): - def getOp(self) -> Incomplete: ... - -class OrExpression(BooleanExpression): - def getOp(self) -> Incomplete: ... - -class ImpExpression(BooleanExpression): - def getOp(self) -> Incomplete: ... - -class IffExpression(BooleanExpression): - def getOp(self) -> Incomplete: ... - -class EqualityExpression(BinaryExpression): - def getOp(self) -> Incomplete: ... - -class LogicalExpressionException(Exception): - index: Incomplete - def __init__(self, index: Incomplete, message: Incomplete) -> None: ... - -class UnexpectedTokenException(LogicalExpressionException): - def __init__( - self, - index: Incomplete, - unexpected: Incomplete | None = None, - expected: Incomplete | None = None, - message: Incomplete | None = None, - ) -> None: ... - -class ExpectedMoreTokensException(LogicalExpressionException): - def __init__( - self, index: Incomplete, message: Incomplete | None = None - ) -> None: ... - -def is_indvar(expr: Incomplete) -> Incomplete: ... -def is_funcvar(expr: Incomplete) -> Incomplete: ... -def is_eventvar(expr: Incomplete) -> Incomplete: ... -def demo() -> None: ... -def demo_errors() -> None: ... -def demoException(s: Incomplete) -> None: ... -def printtype(ex: Incomplete) -> None: ... diff --git a/stubs/nltk/sem/relextract.pyi b/stubs/nltk/sem/relextract.pyi deleted file mode 100644 index f47d3e4..0000000 --- a/stubs/nltk/sem/relextract.pyi +++ /dev/null @@ -1,31 +0,0 @@ -from _typeshed import Incomplete - -NE_CLASSES: Incomplete -short2long: Incomplete -long2short: Incomplete - -def class_abbrev(type: Incomplete) -> Incomplete: ... -def descape_entity(m: Incomplete, defs: Incomplete = ...) -> Incomplete: ... -def list2sym(lst: Incomplete) -> Incomplete: ... -def tree2semi_rel(tree: Incomplete) -> Incomplete: ... -def semi_rel2reldict( - pairs: Incomplete, window: int = 5, trace: bool = False -) -> Incomplete: ... -def extract_rels( - subjclass: Incomplete, - objclass: Incomplete, - doc: Incomplete, - corpus: str = "ace", - pattern: Incomplete | None = None, - window: int = 10, -) -> Incomplete: ... -def rtuple( - reldict: Incomplete, lcon: bool = False, rcon: bool = False -) -> Incomplete: ... -def clause(reldict: Incomplete, relsym: Incomplete) -> Incomplete: ... -def in_demo(trace: int = 0, sql: bool = True) -> None: ... -def roles_demo(trace: int = 0) -> None: ... -def ieer_headlines() -> None: ... -def conllned(trace: int = 1) -> None: ... -def conllesp() -> None: ... -def ne_chunked() -> None: ... diff --git a/stubs/nltk/sem/skolemize.pyi b/stubs/nltk/sem/skolemize.pyi deleted file mode 100644 index 51749c7..0000000 --- a/stubs/nltk/sem/skolemize.pyi +++ /dev/null @@ -1,45 +0,0 @@ -from _typeshed import Incomplete - -from nltk.sem.logic import ( - AllExpression as AllExpression, -) -from nltk.sem.logic import ( - AndExpression as AndExpression, -) -from nltk.sem.logic import ( - ApplicationExpression as ApplicationExpression, -) -from nltk.sem.logic import ( - EqualityExpression as EqualityExpression, -) -from nltk.sem.logic import ( - ExistsExpression as ExistsExpression, -) -from nltk.sem.logic import ( - IffExpression as IffExpression, -) -from nltk.sem.logic import ( - ImpExpression as ImpExpression, -) -from nltk.sem.logic import ( - NegatedExpression as NegatedExpression, -) -from nltk.sem.logic import ( - OrExpression as OrExpression, -) -from nltk.sem.logic import ( - VariableExpression as VariableExpression, -) -from nltk.sem.logic import ( - skolem_function as skolem_function, -) -from nltk.sem.logic import ( - unique_variable as unique_variable, -) - -def skolemize( - expression: Incomplete, - univ_scope: Incomplete | None = None, - used_variables: Incomplete | None = None, -) -> Incomplete: ... -def to_cnf(first: Incomplete, second: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/sem/util.pyi b/stubs/nltk/sem/util.pyi deleted file mode 100644 index aa33ba2..0000000 --- a/stubs/nltk/sem/util.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -from nltk.sem import evaluate as evaluate - -def parse_sents( - inputs: Incomplete, grammar: Incomplete, trace: int = 0 -) -> Incomplete: ... -def root_semrep(syntree: Incomplete, semkey: str = "SEM") -> Incomplete: ... -def interpret_sents( - inputs: Incomplete, - grammar: Incomplete, - semkey: str = "SEM", - trace: int = 0, -) -> Incomplete: ... -def evaluate_sents( - inputs: Incomplete, - grammar: Incomplete, - model: Incomplete, - assignment: Incomplete, - trace: int = 0, -) -> Incomplete: ... -def demo_model0() -> None: ... -def read_sents(filename: Incomplete, encoding: str = "utf8") -> Incomplete: ... -def demo_legacy_grammar() -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/sentiment/__init__.pyi b/stubs/nltk/sentiment/__init__.pyi deleted file mode 100644 index c1c2b96..0000000 --- a/stubs/nltk/sentiment/__init__.pyi +++ /dev/null @@ -1,6 +0,0 @@ -from nltk.sentiment.sentiment_analyzer import ( - SentimentAnalyzer as SentimentAnalyzer, -) -from nltk.sentiment.vader import ( - SentimentIntensityAnalyzer as SentimentIntensityAnalyzer, -) diff --git a/stubs/nltk/sentiment/sentiment_analyzer.pyi b/stubs/nltk/sentiment/sentiment_analyzer.pyi deleted file mode 100644 index 5fbaabb..0000000 --- a/stubs/nltk/sentiment/sentiment_analyzer.pyi +++ /dev/null @@ -1,55 +0,0 @@ -from _typeshed import Incomplete - -from nltk.classify.util import apply_features as apply_features -from nltk.collocations import ( - BigramCollocationFinder as BigramCollocationFinder, -) -from nltk.metrics import BigramAssocMeasures as BigramAssocMeasures -from nltk.probability import FreqDist as FreqDist - -class SentimentAnalyzer: - feat_extractors: Incomplete - classifier: Incomplete - def __init__(self, classifier: Incomplete | None = None) -> None: ... - def all_words( - self, documents: Incomplete, labeled: Incomplete | None = None - ) -> Incomplete: ... - def apply_features( - self, documents: Incomplete, labeled: Incomplete | None = None - ) -> Incomplete: ... - def unigram_word_feats( - self, - words: Incomplete, - top_n: Incomplete | None = None, - min_freq: int = 0, - ) -> Incomplete: ... - def bigram_collocation_feats( - self, - documents: Incomplete, - top_n: Incomplete | None = None, - min_freq: int = 3, - assoc_measure: Incomplete = ..., - ) -> Incomplete: ... - def classify(self, instance: Incomplete) -> Incomplete: ... - def add_feat_extractor( - self, function: Incomplete, **kwargs: Incomplete - ) -> None: ... - def extract_features(self, document: Incomplete) -> Incomplete: ... - def train( - self, - trainer: Incomplete, - training_set: Incomplete, - save_classifier: Incomplete | None = None, - **kwargs: Incomplete, - ) -> Incomplete: ... - def save_file(self, content: Incomplete, filename: Incomplete) -> None: ... - def evaluate( - self, - test_set: Incomplete, - classifier: Incomplete | None = None, - accuracy: bool = True, - f_measure: bool = True, - precision: bool = True, - recall: bool = True, - verbose: bool = False, - ) -> Incomplete: ... diff --git a/stubs/nltk/sentiment/util.pyi b/stubs/nltk/sentiment/util.pyi deleted file mode 100644 index a51ceb7..0000000 --- a/stubs/nltk/sentiment/util.pyi +++ /dev/null @@ -1,73 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus import ( - CategorizedPlaintextCorpusReader as CategorizedPlaintextCorpusReader, -) -from nltk.data import load as load -from nltk.tokenize import PunktTokenizer as PunktTokenizer -from nltk.tokenize.casual import EMOTICON_RE as EMOTICON_RE - -NEGATION: str -NEGATION_RE: Incomplete -CLAUSE_PUNCT: str -CLAUSE_PUNCT_RE: Incomplete -HAPPY: Incomplete -SAD: Incomplete - -def timer(method: Incomplete) -> Incomplete: ... -def extract_unigram_feats( - document: Incomplete, unigrams: Incomplete, handle_negation: bool = False -) -> Incomplete: ... -def extract_bigram_feats( - document: Incomplete, bigrams: Incomplete -) -> Incomplete: ... -def mark_negation( - document: Incomplete, double_neg_flip: bool = False, shallow: bool = False -) -> Incomplete: ... -def output_markdown(filename: Incomplete, **kwargs: Incomplete) -> None: ... -def split_train_test( - all_instances: Incomplete, n: Incomplete | None = None -) -> Incomplete: ... -def json2csv_preprocess( - json_file: Incomplete, - outfile: Incomplete, - fields: Incomplete, - encoding: str = "utf8", - errors: str = "replace", - gzip_compress: bool = False, - skip_retweets: bool = True, - skip_tongue_tweets: bool = True, - skip_ambiguous_tweets: bool = True, - strip_off_emoticons: bool = True, - remove_duplicates: bool = True, - limit: Incomplete | None = None, -) -> None: ... -def parse_tweets_set( - filename: Incomplete, - label: Incomplete, - word_tokenizer: Incomplete | None = None, - sent_tokenizer: Incomplete | None = None, - skip_header: bool = True, -) -> Incomplete: ... -def demo_tweets( - trainer: Incomplete, - n_instances: Incomplete | None = None, - output: Incomplete | None = None, -) -> None: ... -def demo_movie_reviews( - trainer: Incomplete, - n_instances: Incomplete | None = None, - output: Incomplete | None = None, -) -> None: ... -def demo_subjectivity( - trainer: Incomplete, - save_analyzer: bool = False, - n_instances: Incomplete | None = None, - output: Incomplete | None = None, -) -> Incomplete: ... -def demo_sent_subjectivity(text: Incomplete) -> None: ... -def demo_liu_hu_lexicon(sentence: Incomplete, plot: bool = False) -> None: ... -def demo_vader_instance(text: Incomplete) -> None: ... -def demo_vader_tweets( - n_instances: Incomplete | None = None, output: Incomplete | None = None -) -> None: ... diff --git a/stubs/nltk/sentiment/vader.pyi b/stubs/nltk/sentiment/vader.pyi deleted file mode 100644 index e71760d..0000000 --- a/stubs/nltk/sentiment/vader.pyi +++ /dev/null @@ -1,58 +0,0 @@ -from _typeshed import Incomplete - -from nltk.util import pairwise as pairwise - -class VaderConstants: - B_INCR: float - B_DECR: float - C_INCR: float - N_SCALAR: float - NEGATE: Incomplete - BOOSTER_DICT: Incomplete - SPECIAL_CASE_IDIOMS: Incomplete - REGEX_REMOVE_PUNCTUATION: Incomplete - PUNC_LIST: Incomplete - def __init__(self) -> None: ... - def negated( - self, input_words: Incomplete, include_nt: bool = True - ) -> Incomplete: ... - def normalize(self, score: Incomplete, alpha: int = 15) -> Incomplete: ... - def scalar_inc_dec( - self, word: Incomplete, valence: Incomplete, is_cap_diff: Incomplete - ) -> Incomplete: ... - -class SentiText: - text: Incomplete - PUNC_LIST: Incomplete - REGEX_REMOVE_PUNCTUATION: Incomplete - words_and_emoticons: Incomplete - is_cap_diff: Incomplete - def __init__( - self, - text: Incomplete, - punc_list: Incomplete, - regex_remove_punctuation: Incomplete, - ) -> None: ... - def allcap_differential(self, words: Incomplete) -> Incomplete: ... - -class SentimentIntensityAnalyzer: - lexicon_file: Incomplete - lexicon: Incomplete - constants: Incomplete - def __init__( - self, - lexicon_file: str = "sentiment/vader_lexicon.zip/vader_lexicon/vader_lexicon.txt", - ) -> None: ... - def make_lex_dict(self) -> Incomplete: ... - def polarity_scores(self, text: Incomplete) -> Incomplete: ... - def sentiment_valence( - self, - valence: Incomplete, - sentitext: Incomplete, - item: Incomplete, - i: Incomplete, - sentiments: Incomplete, - ) -> Incomplete: ... - def score_valence( - self, sentiments: Incomplete, text: Incomplete - ) -> Incomplete: ... diff --git a/stubs/nltk/stem/__init__.pyi b/stubs/nltk/stem/__init__.pyi deleted file mode 100644 index 1482d3d..0000000 --- a/stubs/nltk/stem/__init__.pyi +++ /dev/null @@ -1,11 +0,0 @@ -from nltk.stem.api import StemmerI as StemmerI -from nltk.stem.arlstem import ARLSTem as ARLSTem -from nltk.stem.arlstem2 import ARLSTem2 as ARLSTem2 -from nltk.stem.cistem import Cistem as Cistem -from nltk.stem.isri import ISRIStemmer as ISRIStemmer -from nltk.stem.lancaster import LancasterStemmer as LancasterStemmer -from nltk.stem.porter import PorterStemmer as PorterStemmer -from nltk.stem.regexp import RegexpStemmer as RegexpStemmer -from nltk.stem.rslp import RSLPStemmer as RSLPStemmer -from nltk.stem.snowball import SnowballStemmer as SnowballStemmer -from nltk.stem.wordnet import WordNetLemmatizer as WordNetLemmatizer diff --git a/stubs/nltk/stem/api.pyi b/stubs/nltk/stem/api.pyi deleted file mode 100644 index 69d9b7e..0000000 --- a/stubs/nltk/stem/api.pyi +++ /dev/null @@ -1,7 +0,0 @@ -from abc import ABCMeta, abstractmethod - -from _typeshed import Incomplete - -class StemmerI(metaclass=ABCMeta): - @abstractmethod - def stem(self, token: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/stem/arlstem.pyi b/stubs/nltk/stem/arlstem.pyi deleted file mode 100644 index 92ba5b3..0000000 --- a/stubs/nltk/stem/arlstem.pyi +++ /dev/null @@ -1,39 +0,0 @@ -from _typeshed import Incomplete - -from nltk.stem.api import StemmerI as StemmerI - -class ARLSTem(StemmerI): - re_hamzated_alif: Incomplete - re_alifMaqsura: Incomplete - re_diacritics: Incomplete - pr2: Incomplete - pr3: Incomplete - pr32: Incomplete - pr4: Incomplete - su2: Incomplete - su22: Incomplete - su3: Incomplete - su32: Incomplete - pl_si2: Incomplete - pl_si3: Incomplete - verb_su2: Incomplete - verb_pr2: Incomplete - verb_pr22: Incomplete - verb_pr33: Incomplete - verb_suf3: Incomplete - verb_suf2: Incomplete - verb_suf1: Incomplete - def __init__(self) -> None: ... - def stem(self, token: Incomplete) -> Incomplete: ... - def norm(self, token: Incomplete) -> Incomplete: ... - def pref(self, token: Incomplete) -> Incomplete: ... - def suff(self, token: Incomplete) -> Incomplete: ... - def fem2masc(self, token: Incomplete) -> Incomplete: ... - def plur2sing(self, token: Incomplete) -> Incomplete: ... - def verb(self, token: Incomplete) -> Incomplete: ... - def verb_t1(self, token: Incomplete) -> Incomplete: ... - def verb_t2(self, token: Incomplete) -> Incomplete: ... - def verb_t3(self, token: Incomplete) -> Incomplete: ... - def verb_t4(self, token: Incomplete) -> Incomplete: ... - def verb_t5(self, token: Incomplete) -> Incomplete: ... - def verb_t6(self, token: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/stem/arlstem2.pyi b/stubs/nltk/stem/arlstem2.pyi deleted file mode 100644 index 76bbeef..0000000 --- a/stubs/nltk/stem/arlstem2.pyi +++ /dev/null @@ -1,42 +0,0 @@ -from _typeshed import Incomplete - -from nltk.stem.api import StemmerI as StemmerI - -class ARLSTem2(StemmerI): - re_hamzated_alif: Incomplete - re_alifMaqsura: Incomplete - re_diacritics: Incomplete - pr2: Incomplete - pr3: Incomplete - pr32: Incomplete - pr4: Incomplete - su2: Incomplete - su22: Incomplete - su3: Incomplete - su32: Incomplete - pl_si2: Incomplete - pl_si3: Incomplete - verb_su2: Incomplete - verb_pr2: Incomplete - verb_pr22: Incomplete - verb_pr33: Incomplete - verb_suf3: Incomplete - verb_suf2: Incomplete - verb_suf1: Incomplete - def __init__(self) -> None: ... - is_verb: bool - def stem1(self, token: Incomplete) -> Incomplete: ... - def stem(self, token: Incomplete) -> Incomplete: ... - def norm(self, token: Incomplete) -> Incomplete: ... - def pref(self, token: Incomplete) -> Incomplete: ... - def adjective(self, token: Incomplete) -> Incomplete: ... - def suff(self, token: Incomplete) -> Incomplete: ... - def fem2masc(self, token: Incomplete) -> Incomplete: ... - def plur2sing(self, token: Incomplete) -> Incomplete: ... - def verb(self, token: Incomplete) -> Incomplete: ... - def verb_t1(self, token: Incomplete) -> Incomplete: ... - def verb_t2(self, token: Incomplete) -> Incomplete: ... - def verb_t3(self, token: Incomplete) -> Incomplete: ... - def verb_t4(self, token: Incomplete) -> Incomplete: ... - def verb_t5(self, token: Incomplete) -> Incomplete: ... - def verb_t6(self, token: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/stem/cistem.pyi b/stubs/nltk/stem/cistem.pyi deleted file mode 100644 index a524151..0000000 --- a/stubs/nltk/stem/cistem.pyi +++ /dev/null @@ -1,19 +0,0 @@ -from _typeshed import Incomplete - -from nltk.stem.api import StemmerI as StemmerI - -class Cistem(StemmerI): - strip_ge: Incomplete - repl_xx: Incomplete - strip_emr: Incomplete - strip_nd: Incomplete - strip_t: Incomplete - strip_esn: Incomplete - repl_xx_back: Incomplete - def __init__(self, case_insensitive: bool = False) -> None: ... - @staticmethod - def replace_to(word: str) -> str: ... - @staticmethod - def replace_back(word: str) -> str: ... - def stem(self, word: str) -> str: ... - def segment(self, word: str) -> tuple[str, str]: ... diff --git a/stubs/nltk/stem/isri.pyi b/stubs/nltk/stem/isri.pyi deleted file mode 100644 index 2a04a25..0000000 --- a/stubs/nltk/stem/isri.pyi +++ /dev/null @@ -1,32 +0,0 @@ -from _typeshed import Incomplete - -from nltk.stem.api import StemmerI as StemmerI - -class ISRIStemmer(StemmerI): - p3: Incomplete - p2: Incomplete - p1: Incomplete - s3: Incomplete - s2: Incomplete - s1: Incomplete - pr4: Incomplete - pr53: Incomplete - re_short_vowels: Incomplete - re_hamza: Incomplete - re_initial_hamza: Incomplete - stop_words: Incomplete - def __init__(self) -> None: ... - def stem(self, token: Incomplete) -> Incomplete: ... - def norm(self, word: Incomplete, num: int = 3) -> Incomplete: ... - def pre32(self, word: Incomplete) -> Incomplete: ... - def suf32(self, word: Incomplete) -> Incomplete: ... - def waw(self, word: Incomplete) -> Incomplete: ... - def pro_w4(self, word: Incomplete) -> Incomplete: ... - def pro_w53(self, word: Incomplete) -> Incomplete: ... - def pro_w54(self, word: Incomplete) -> Incomplete: ... - def end_w5(self, word: Incomplete) -> Incomplete: ... - def pro_w6(self, word: Incomplete) -> Incomplete: ... - def pro_w64(self, word: Incomplete) -> Incomplete: ... - def end_w6(self, word: Incomplete) -> Incomplete: ... - def suf1(self, word: Incomplete) -> Incomplete: ... - def pre1(self, word: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/stem/lancaster.pyi b/stubs/nltk/stem/lancaster.pyi deleted file mode 100644 index 994f15a..0000000 --- a/stubs/nltk/stem/lancaster.pyi +++ /dev/null @@ -1,14 +0,0 @@ -from _typeshed import Incomplete - -from nltk.stem.api import StemmerI as StemmerI - -class LancasterStemmer(StemmerI): - default_rule_tuple: Incomplete - rule_dictionary: Incomplete - def __init__( - self, - rule_tuple: Incomplete | None = None, - strip_prefix_flag: bool = False, - ) -> None: ... - def parseRules(self, rule_tuple: Incomplete | None = None) -> None: ... - def stem(self, word: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/stem/porter.pyi b/stubs/nltk/stem/porter.pyi deleted file mode 100644 index 0aa44fb..0000000 --- a/stubs/nltk/stem/porter.pyi +++ /dev/null @@ -1,19 +0,0 @@ -from _typeshed import Incomplete - -from nltk.stem.api import StemmerI as StemmerI - -__docformat__: str - -class PorterStemmer(StemmerI): - NLTK_EXTENSIONS: str - MARTIN_EXTENSIONS: str - ORIGINAL_ALGORITHM: str - mode: Incomplete - pool: Incomplete - vowels: Incomplete - def __init__(self, mode: Incomplete = ...) -> None: ... - def stem( - self, word: Incomplete, to_lowercase: bool = True - ) -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/stem/regexp.pyi b/stubs/nltk/stem/regexp.pyi deleted file mode 100644 index d2bd90f..0000000 --- a/stubs/nltk/stem/regexp.pyi +++ /dev/null @@ -1,7 +0,0 @@ -from _typeshed import Incomplete - -from nltk.stem.api import StemmerI as StemmerI - -class RegexpStemmer(StemmerI): - def __init__(self, regexp: Incomplete, min: int = 0) -> None: ... - def stem(self, word: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/stem/rslp.pyi b/stubs/nltk/stem/rslp.pyi deleted file mode 100644 index 6610fc6..0000000 --- a/stubs/nltk/stem/rslp.pyi +++ /dev/null @@ -1,12 +0,0 @@ -from _typeshed import Incomplete - -from nltk.data import load as load -from nltk.stem.api import StemmerI as StemmerI - -class RSLPStemmer(StemmerI): - def __init__(self) -> None: ... - def read_rule(self, filename: Incomplete) -> Incomplete: ... - def stem(self, word: Incomplete) -> Incomplete: ... - def apply_rule( - self, word: Incomplete, rule_index: Incomplete - ) -> Incomplete: ... diff --git a/stubs/nltk/stem/snowball.pyi b/stubs/nltk/stem/snowball.pyi deleted file mode 100644 index dd04bff..0000000 --- a/stubs/nltk/stem/snowball.pyi +++ /dev/null @@ -1,91 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus import stopwords as stopwords -from nltk.stem import porter as porter -from nltk.stem.api import StemmerI as StemmerI -from nltk.stem.util import ( - prefix_replace as prefix_replace, -) -from nltk.stem.util import ( - suffix_replace as suffix_replace, -) - -class SnowballStemmer(StemmerI): - languages: Incomplete - stemmer: Incomplete - stopwords: Incomplete - def __init__( - self, language: Incomplete, ignore_stopwords: bool = False - ) -> None: ... - def stem(self, token: Incomplete) -> Incomplete: ... - -class _LanguageSpecificStemmer(StemmerI): # type: ignore[misc] - stopwords: Incomplete - def __init__(self, ignore_stopwords: bool = False) -> None: ... - -class PorterStemmer(_LanguageSpecificStemmer, porter.PorterStemmer): - def __init__(self, ignore_stopwords: bool = False) -> None: ... - -class _ScandinavianStemmer(_LanguageSpecificStemmer): ... # type: ignore[misc] -class _StandardStemmer(_LanguageSpecificStemmer): ... - -class ArabicStemmer(_StandardStemmer): - is_verb: bool - is_noun: bool - is_defined: bool - suffixes_verb_step1_success: bool - suffix_verb_step2a_success: bool - suffix_verb_step2b_success: bool - suffix_noun_step2c2_success: bool - suffix_noun_step1a_success: bool - suffix_noun_step2a_success: bool - suffix_noun_step2b_success: bool - suffixe_noun_step1b_success: bool - prefix_step2a_success: bool - prefix_step3a_noun_success: bool - prefix_step3b_noun_success: bool - def stem(self, word: Incomplete) -> Incomplete: ... - -class DanishStemmer(_ScandinavianStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class DutchStemmer(_StandardStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class EnglishStemmer(_StandardStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class FinnishStemmer(_StandardStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class FrenchStemmer(_StandardStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class GermanStemmer(_StandardStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class HungarianStemmer(_LanguageSpecificStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class ItalianStemmer(_StandardStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class NorwegianStemmer(_ScandinavianStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class PortugueseStemmer(_StandardStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class RomanianStemmer(_StandardStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class RussianStemmer(_LanguageSpecificStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class SpanishStemmer(_StandardStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -class SwedishStemmer(_ScandinavianStemmer): - def stem(self, word: Incomplete) -> Incomplete: ... - -def demo() -> None: ... diff --git a/stubs/nltk/stem/util.pyi b/stubs/nltk/stem/util.pyi deleted file mode 100644 index 90713e2..0000000 --- a/stubs/nltk/stem/util.pyi +++ /dev/null @@ -1,8 +0,0 @@ -from _typeshed import Incomplete - -def suffix_replace( - original: Incomplete, old: Incomplete, new: Incomplete -) -> Incomplete: ... -def prefix_replace( - original: Incomplete, old: Incomplete, new: Incomplete -) -> Incomplete: ... diff --git a/stubs/nltk/stem/wordnet.pyi b/stubs/nltk/stem/wordnet.pyi deleted file mode 100644 index 8444b7b..0000000 --- a/stubs/nltk/stem/wordnet.pyi +++ /dev/null @@ -1,10 +0,0 @@ -from _typeshed import Incomplete - -class WordNetLemmatizer: - def morphy( - self, - form: Incomplete, - pos: Incomplete | None = None, - check_exceptions: bool = True, - ) -> Incomplete: ... - def lemmatize(self, word: str, pos: str = "n") -> str: ... diff --git a/stubs/nltk/tabdata.pyi b/stubs/nltk/tabdata.pyi deleted file mode 100644 index 8503c21..0000000 --- a/stubs/nltk/tabdata.pyi +++ /dev/null @@ -1,28 +0,0 @@ -from _typeshed import Incomplete - -def rm_nl(s: Incomplete) -> Incomplete: ... - -class TabEncoder: - def list2txt(self, s: Incomplete) -> Incomplete: ... - def set2txt(self, s: Incomplete) -> Incomplete: ... - def tup2tab(self, tup: Incomplete) -> Incomplete: ... - def tups2tab(self, x: Incomplete) -> Incomplete: ... - def dict2tab(self, d: Incomplete) -> Incomplete: ... - def ivdict2tab(self, d: Incomplete) -> Incomplete: ... - -class TabDecoder: - def txt2list(self, f: Incomplete) -> Incomplete: ... - def txt2set(self, f: Incomplete) -> Incomplete: ... - def tab2tup(self, s: Incomplete) -> Incomplete: ... - def tab2tups(self, f: Incomplete) -> Incomplete: ... - def tab2dict(self, f: Incomplete) -> Incomplete: ... - def tab2ivdict(self, f: Incomplete) -> Incomplete: ... - -class MaxentEncoder(TabEncoder): - def tupdict2tab(self, d: Incomplete) -> Incomplete: ... - -class MaxentDecoder(TabDecoder): - def tupkey2dict(self, f: Incomplete) -> Incomplete: ... - -class PunktDecoder(TabDecoder): - def tab2intdict(self, f: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tag/__init__.pyi b/stubs/nltk/tag/__init__.pyi deleted file mode 100644 index ab0a8ac..0000000 --- a/stubs/nltk/tag/__init__.pyi +++ /dev/null @@ -1,92 +0,0 @@ -from _typeshed import Incomplete - -from nltk.data import find as find -from nltk.data import load as load -from nltk.tag.api import TaggerI as TaggerI -from nltk.tag.brill import BrillTagger as BrillTagger -from nltk.tag.brill_trainer import BrillTaggerTrainer as BrillTaggerTrainer -from nltk.tag.crf import CRFTagger as CRFTagger -from nltk.tag.hmm import ( - HiddenMarkovModelTagger as HiddenMarkovModelTagger, -) -from nltk.tag.hmm import ( - HiddenMarkovModelTrainer as HiddenMarkovModelTrainer, -) -from nltk.tag.hunpos import HunposTagger as HunposTagger -from nltk.tag.mapping import ( - map_tag as map_tag, -) -from nltk.tag.mapping import ( - tagset_mapping as tagset_mapping, -) -from nltk.tag.perceptron import PerceptronTagger as PerceptronTagger -from nltk.tag.senna import ( - SennaChunkTagger as SennaChunkTagger, -) -from nltk.tag.senna import ( - SennaNERTagger as SennaNERTagger, -) -from nltk.tag.senna import ( - SennaTagger as SennaTagger, -) -from nltk.tag.sequential import ( - AffixTagger as AffixTagger, -) -from nltk.tag.sequential import ( - BigramTagger as BigramTagger, -) -from nltk.tag.sequential import ( - ClassifierBasedPOSTagger as ClassifierBasedPOSTagger, -) -from nltk.tag.sequential import ( - ClassifierBasedTagger as ClassifierBasedTagger, -) -from nltk.tag.sequential import ( - ContextTagger as ContextTagger, -) -from nltk.tag.sequential import ( - DefaultTagger as DefaultTagger, -) -from nltk.tag.sequential import ( - NgramTagger as NgramTagger, -) -from nltk.tag.sequential import ( - RegexpTagger as RegexpTagger, -) -from nltk.tag.sequential import ( - SequentialBackoffTagger as SequentialBackoffTagger, -) -from nltk.tag.sequential import ( - TrigramTagger as TrigramTagger, -) -from nltk.tag.sequential import ( - UnigramTagger as UnigramTagger, -) -from nltk.tag.stanford import ( - StanfordNERTagger as StanfordNERTagger, -) -from nltk.tag.stanford import ( - StanfordPOSTagger as StanfordPOSTagger, -) -from nltk.tag.stanford import ( - StanfordTagger as StanfordTagger, -) -from nltk.tag.tnt import TnT as TnT -from nltk.tag.util import ( - str2tuple as str2tuple, -) -from nltk.tag.util import ( - tuple2str as tuple2str, -) -from nltk.tag.util import ( - untag as untag, -) - -PRETRAINED_TAGGERS: Incomplete - -def pos_tag( - tokens: Incomplete, tagset: Incomplete | None = None, lang: str = "eng" -) -> Incomplete: ... -def pos_tag_sents( - sentences: Incomplete, tagset: Incomplete | None = None, lang: str = "eng" -) -> Incomplete: ... diff --git a/stubs/nltk/tag/api.pyi b/stubs/nltk/tag/api.pyi deleted file mode 100644 index 24f0148..0000000 --- a/stubs/nltk/tag/api.pyi +++ /dev/null @@ -1,35 +0,0 @@ -from abc import ABCMeta, abstractmethod - -from _typeshed import Incomplete - -from nltk.internals import deprecated as deprecated -from nltk.internals import overridden as overridden -from nltk.metrics import ( - ConfusionMatrix as ConfusionMatrix, -) -from nltk.metrics import ( - accuracy as accuracy, -) -from nltk.tag.util import untag as untag - -class TaggerI(metaclass=ABCMeta): - @abstractmethod - def tag(self, tokens: Incomplete) -> Incomplete: ... - def tag_sents(self, sentences: Incomplete) -> Incomplete: ... - def evaluate(self, gold: Incomplete) -> Incomplete: ... - def accuracy(self, gold: Incomplete) -> Incomplete: ... - def confusion(self, gold: Incomplete) -> Incomplete: ... - def recall(self, gold: Incomplete) -> dict[str, float]: ... - def precision(self, gold: Incomplete) -> Incomplete: ... - def f_measure( - self, gold: Incomplete, alpha: float = 0.5 - ) -> Incomplete: ... - def evaluate_per_tag( - self, - gold: Incomplete, - alpha: float = 0.5, - truncate: Incomplete | None = None, - sort_by_count: bool = False, - ) -> Incomplete: ... - -class FeaturesetTaggerI(TaggerI): ... diff --git a/stubs/nltk/tag/brill.pyi b/stubs/nltk/tag/brill.pyi deleted file mode 100644 index c9fc0ef..0000000 --- a/stubs/nltk/tag/brill.pyi +++ /dev/null @@ -1,49 +0,0 @@ -from _typeshed import Incomplete - -from nltk import jsontags as jsontags -from nltk.tag import TaggerI as TaggerI -from nltk.tbl import Feature as Feature -from nltk.tbl import Template as Template - -class Word(Feature): - json_tag: str - @staticmethod - def extract_property( - tokens: Incomplete, index: Incomplete - ) -> Incomplete: ... - -class Pos(Feature): - json_tag: str - @staticmethod - def extract_property( - tokens: Incomplete, index: Incomplete - ) -> Incomplete: ... - -def nltkdemo18() -> Incomplete: ... -def nltkdemo18plus() -> Incomplete: ... -def fntbl37() -> Incomplete: ... -def brill24() -> Incomplete: ... -def describe_template_sets() -> None: ... - -class BrillTagger(TaggerI): - json_tag: str - def __init__( - self, - initial_tagger: Incomplete, - rules: Incomplete, - training_stats: Incomplete | None = None, - ) -> None: ... - def encode_json_obj(self) -> Incomplete: ... - @classmethod - def decode_json_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... - def rules(self) -> Incomplete: ... - def train_stats( - self, statistic: Incomplete | None = None - ) -> Incomplete: ... - def tag(self, tokens: Incomplete) -> Incomplete: ... - def print_template_statistics( - self, test_stats: Incomplete | None = None, printunused: bool = True - ) -> Incomplete: ... - def batch_tag_incremental( - self, sequences: Incomplete, gold: Incomplete - ) -> Incomplete: ... diff --git a/stubs/nltk/tag/brill_trainer.pyi b/stubs/nltk/tag/brill_trainer.pyi deleted file mode 100644 index 116b08c..0000000 --- a/stubs/nltk/tag/brill_trainer.pyi +++ /dev/null @@ -1,21 +0,0 @@ -from _typeshed import Incomplete - -from nltk.tag import BrillTagger as BrillTagger -from nltk.tag import untag as untag - -class BrillTaggerTrainer: - def __init__( - self, - initial_tagger: Incomplete, - templates: Incomplete, - trace: int = 0, - deterministic: Incomplete | None = None, - ruleformat: str = "str", - ) -> None: ... - def train( - self, - train_sents: Incomplete, - max_rules: int = 200, - min_score: int = 2, - min_acc: Incomplete | None = None, - ) -> Incomplete: ... diff --git a/stubs/nltk/tag/crf.pyi b/stubs/nltk/tag/crf.pyi deleted file mode 100644 index 9ee0777..0000000 --- a/stubs/nltk/tag/crf.pyi +++ /dev/null @@ -1,17 +0,0 @@ -from _typeshed import Incomplete - -from nltk.tag.api import TaggerI as TaggerI - -class CRFTagger(TaggerI): - def __init__( - self, - feature_func: Incomplete | None = None, - verbose: bool = False, - training_opt: Incomplete = {}, - ) -> None: ... - def set_model_file(self, model_file: Incomplete) -> None: ... - def tag_sents(self, sents: Incomplete) -> Incomplete: ... - def train( - self, train_data: Incomplete, model_file: Incomplete - ) -> None: ... - def tag(self, tokens: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tag/hmm.pyi b/stubs/nltk/tag/hmm.pyi deleted file mode 100644 index 422c438..0000000 --- a/stubs/nltk/tag/hmm.pyi +++ /dev/null @@ -1,138 +0,0 @@ -from typing import ( - Callable, - List, - Tuple, - Union, -) - -from _typeshed import Incomplete -from numpy import ( - float64, - ndarray, -) - -from nltk.metrics import accuracy as accuracy -from nltk.probability import ( - ConditionalFreqDist as ConditionalFreqDist, -) -from nltk.probability import ( - ConditionalProbDist as ConditionalProbDist, -) -from nltk.probability import ( - DictionaryConditionalProbDist, - DictionaryProbDist, -) -from nltk.probability import ( - FreqDist as FreqDist, -) -from nltk.probability import ( - LidstoneProbDist as LidstoneProbDist, -) -from nltk.probability import ( - MLEProbDist as MLEProbDist, -) -from nltk.probability import ( - MutableProbDist as MutableProbDist, -) -from nltk.probability import ( - RandomProbDist as RandomProbDist, -) -from nltk.tag.api import TaggerI as TaggerI -from nltk.util import LazyMap as LazyMap -from nltk.util import unique_list as unique_list - -def _create_hmm_tagger( - states: List[str], - symbols: List[str], - A: Union[List[List[float]], ndarray], # type: ignore[type-arg] - B: Union[List[List[float]], ndarray], # type: ignore[type-arg] - pi: Union[List[float], ndarray], # type: ignore[type-arg] -) -> HiddenMarkovModelTagger: ... -def _market_hmm_example() -> Tuple[ - HiddenMarkovModelTagger, List[str], List[str] -]: ... -def _ninf_array(shape: Tuple[int, int]) -> ndarray: ... # type: ignore[type-arg] -def logsumexp2(arr: ndarray) -> float64: ... # type: ignore[type-arg] - -class HiddenMarkovModelTagger(TaggerI): - def __init__( - self, - symbols: List[str], - states: List[str], - transitions: DictionaryConditionalProbDist, - outputs: DictionaryConditionalProbDist, - priors: DictionaryProbDist, - transform: Callable = ..., # type: ignore[type-arg] - ) -> None: ... - def _backward_probability( - self, unlabeled_sequence: List[Tuple[str, None]] - ) -> ndarray: ... - def _forward_probability( - self, unlabeled_sequence: List[Tuple[str, None]] - ) -> ndarray: ... - def _output_logprob(self, state: str, symbol: str) -> float: ... - def _outputs_vector(self, symbol: str) -> ndarray: ... # type: ignore[type-arg] - def _transitions_matrix(self) -> ndarray: ... # type: ignore[type-arg] - @classmethod - def train( - cls: Incomplete, - labeled_sequence: Incomplete, - test_sequence: Incomplete | None = None, - unlabeled_sequence: Incomplete | None = None, - **kwargs: Incomplete, - ) -> Incomplete: ... - def probability(self, sequence: Incomplete) -> Incomplete: ... - def log_probability(self, sequence: Incomplete) -> Incomplete: ... - def tag(self, unlabeled_sequence: Incomplete) -> Incomplete: ... - def reset_cache(self) -> None: ... - def best_path(self, unlabeled_sequence: Incomplete) -> Incomplete: ... - def best_path_simple( - self, unlabeled_sequence: Incomplete - ) -> Incomplete: ... - def random_sample( - self, rng: Incomplete, length: Incomplete - ) -> Incomplete: ... - def entropy(self, unlabeled_sequence: Incomplete) -> Incomplete: ... - def point_entropy(self, unlabeled_sequence: Incomplete) -> Incomplete: ... - def test( - self, - test_sequence: Incomplete, - verbose: bool = False, - **kwargs: Incomplete, - ) -> Incomplete: ... - -class HiddenMarkovModelTrainer: - def __init__( - self, - states: Incomplete | None = None, - symbols: Incomplete | None = None, - ) -> None: ... - def train( - self, - labeled_sequences: Incomplete | None = None, - unlabeled_sequences: Incomplete | None = None, - **kwargs: Incomplete, - ) -> Incomplete: ... - def train_unsupervised( - self, - unlabeled_sequences: Incomplete, - update_outputs: bool = True, - **kwargs: Incomplete, - ) -> Incomplete: ... - def train_supervised( - self, - labelled_sequences: Incomplete, - estimator: Incomplete | None = None, - ) -> Incomplete: ... - -def demo() -> None: ... -def load_pos(num_sents: Incomplete) -> Incomplete: ... -def demo_pos() -> Incomplete: ... -def demo_pos_bw( - test: int = 10, - supervised: int = 20, - unsupervised: int = 10, - verbose: bool = True, - max_iterations: int = 5, -) -> Incomplete: ... -def demo_bw() -> None: ... diff --git a/stubs/nltk/tag/hunpos.pyi b/stubs/nltk/tag/hunpos.pyi deleted file mode 100644 index 26a5224..0000000 --- a/stubs/nltk/tag/hunpos.pyi +++ /dev/null @@ -1,26 +0,0 @@ -import types - -from _typeshed import Incomplete - -from nltk.internals import find_binary as find_binary -from nltk.internals import find_file as find_file -from nltk.tag.api import TaggerI as TaggerI - -class HunposTagger(TaggerI): - def __init__( - self, - path_to_model: Incomplete, - path_to_bin: Incomplete | None = None, - encoding: Incomplete = ..., - verbose: bool = False, - ) -> None: ... - def __del__(self) -> None: ... - def close(self) -> None: ... - def __enter__(self) -> Incomplete: ... - def __exit__( - self, - exc_type: type[BaseException] | None, - exc_value: BaseException | None, - traceback: types.TracebackType | None, - ) -> None: ... - def tag(self, tokens: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tag/mapping.pyi b/stubs/nltk/tag/mapping.pyi deleted file mode 100644 index 5f2238c..0000000 --- a/stubs/nltk/tag/mapping.pyi +++ /dev/null @@ -1,8 +0,0 @@ -from _typeshed import Incomplete - -from nltk.data import load as load - -def tagset_mapping(source: Incomplete, target: Incomplete) -> Incomplete: ... -def map_tag( - source: Incomplete, target: Incomplete, source_tag: Incomplete -) -> Incomplete: ... diff --git a/stubs/nltk/tag/perceptron.pyi b/stubs/nltk/tag/perceptron.pyi deleted file mode 100644 index ed0bea7..0000000 --- a/stubs/nltk/tag/perceptron.pyi +++ /dev/null @@ -1,55 +0,0 @@ -from _typeshed import Incomplete - -from nltk import jsontags as jsontags -from nltk.data import find as find -from nltk.data import load as load -from nltk.tag.api import TaggerI as TaggerI - -TRAINED_TAGGER_PATH: str -TAGGER_JSONS: Incomplete - -class AveragedPerceptron: - json_tag: str - weights: Incomplete - classes: Incomplete - i: int - def __init__(self, weights: Incomplete | None = None) -> None: ... - def predict( - self, features: Incomplete, return_conf: bool = False - ) -> Incomplete: ... - def update( - self, truth: Incomplete, guess: Incomplete, features: Incomplete - ) -> None: ... - def average_weights(self) -> None: ... - def save(self, path: Incomplete) -> Incomplete: ... - def load(self, path: Incomplete) -> None: ... - def encode_json_obj(self) -> Incomplete: ... - @classmethod - def decode_json_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... - -class PerceptronTagger(TaggerI): - json_tag: str - START: Incomplete - END: Incomplete - model: Incomplete - tagdict: Incomplete - classes: Incomplete - def __init__(self, load: bool = True, lang: str = "eng") -> None: ... - def tag( - self, - tokens: Incomplete, - return_conf: bool = False, - use_tagdict: bool = True, - ) -> Incomplete: ... - def train( - self, - sentences: Incomplete, - save_loc: Incomplete | None = None, - nr_iter: int = 5, - ) -> None: ... - def save_to_json(self, loc: Incomplete, lang: str = "xxx") -> None: ... - def load_from_json(self, lang: str = "eng") -> None: ... - def encode_json_obj(self) -> Incomplete: ... - @classmethod - def decode_json_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... - def normalize(self, word: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tag/senna.pyi b/stubs/nltk/tag/senna.pyi deleted file mode 100644 index 9d49772..0000000 --- a/stubs/nltk/tag/senna.pyi +++ /dev/null @@ -1,20 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.classify import Senna as Senna - -class SennaTagger(Senna): - def __init__(self, path: Incomplete, encoding: str = "utf-8") -> None: ... - def tag_sents(self, sentences: Incomplete) -> Incomplete: ... - -class SennaChunkTagger(Senna): - def __init__(self, path: Incomplete, encoding: str = "utf-8") -> None: ... - def tag_sents(self, sentences: Incomplete) -> Incomplete: ... - def bio_to_chunks( - self, tagged_sent: Incomplete, chunk_type: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class SennaNERTagger(Senna): - def __init__(self, path: Incomplete, encoding: str = "utf-8") -> None: ... - def tag_sents(self, sentences: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tag/sequential.pyi b/stubs/nltk/tag/sequential.pyi deleted file mode 100644 index 065d669..0000000 --- a/stubs/nltk/tag/sequential.pyi +++ /dev/null @@ -1,158 +0,0 @@ -from abc import abstractmethod - -from _typeshed import Incomplete - -from nltk import jsontags as jsontags -from nltk.classify import NaiveBayesClassifier as NaiveBayesClassifier -from nltk.probability import ConditionalFreqDist as ConditionalFreqDist -from nltk.tag.api import ( - FeaturesetTaggerI as FeaturesetTaggerI, -) -from nltk.tag.api import ( - TaggerI as TaggerI, -) - -class SequentialBackoffTagger(TaggerI): - def __init__(self, backoff: Incomplete | None = None) -> None: ... - @property - def backoff(self) -> Incomplete: ... - def tag(self, tokens: Incomplete) -> Incomplete: ... - def tag_one( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - @abstractmethod - def choose_tag( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - -class ContextTagger(SequentialBackoffTagger): - def __init__( - self, context_to_tag: Incomplete, backoff: Incomplete | None = None - ) -> None: ... - @abstractmethod - def context( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - def choose_tag( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - def size(self) -> Incomplete: ... - -class DefaultTagger(SequentialBackoffTagger): - json_tag: str - def __init__(self, tag: Incomplete) -> None: ... - def encode_json_obj(self) -> Incomplete: ... - @classmethod - def decode_json_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... - def choose_tag( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - -class NgramTagger(ContextTagger): - json_tag: str - def __init__( - self, - n: Incomplete, - train: Incomplete | None = None, - model: Incomplete | None = None, - backoff: Incomplete | None = None, - cutoff: int = 0, - verbose: bool = False, - ) -> None: ... - def encode_json_obj(self) -> Incomplete: ... - @classmethod - def decode_json_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... - def context( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - -class UnigramTagger(NgramTagger): - json_tag: str - def __init__( - self, - train: Incomplete | None = None, - model: Incomplete | None = None, - backoff: Incomplete | None = None, - cutoff: int = 0, - verbose: bool = False, - ) -> None: ... - def context( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - -class BigramTagger(NgramTagger): - json_tag: str - def __init__( - self, - train: Incomplete | None = None, - model: Incomplete | None = None, - backoff: Incomplete | None = None, - cutoff: int = 0, - verbose: bool = False, - ) -> None: ... - -class TrigramTagger(NgramTagger): - json_tag: str - def __init__( - self, - train: Incomplete | None = None, - model: Incomplete | None = None, - backoff: Incomplete | None = None, - cutoff: int = 0, - verbose: bool = False, - ) -> None: ... - -class AffixTagger(ContextTagger): - json_tag: str - def __init__( - self, - train: Incomplete | None = None, - model: Incomplete | None = None, - affix_length: int = -3, - min_stem_length: int = 2, - backoff: Incomplete | None = None, - cutoff: int = 0, - verbose: bool = False, - ) -> None: ... - def encode_json_obj(self) -> Incomplete: ... - @classmethod - def decode_json_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... - def context( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - -class RegexpTagger(SequentialBackoffTagger): - json_tag: str - def __init__( - self, regexps: list[tuple[str, str]], backoff: TaggerI | None = None - ) -> None: ... - def encode_json_obj(self) -> Incomplete: ... - @classmethod - def decode_json_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... - def choose_tag( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - -class ClassifierBasedTagger(SequentialBackoffTagger, FeaturesetTaggerI): - def __init__( - self, - feature_detector: Incomplete | None = None, - train: Incomplete | None = None, - classifier_builder: Incomplete = ..., - classifier: Incomplete | None = None, - backoff: Incomplete | None = None, - cutoff_prob: Incomplete | None = None, - verbose: bool = False, - ) -> None: ... - def choose_tag( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - def feature_detector( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... - def classifier(self) -> Incomplete: ... - -class ClassifierBasedPOSTagger(ClassifierBasedTagger): - def feature_detector( - self, tokens: Incomplete, index: Incomplete, history: Incomplete - ) -> Incomplete: ... diff --git a/stubs/nltk/tag/stanford.pyi b/stubs/nltk/tag/stanford.pyi deleted file mode 100644 index baaaff2..0000000 --- a/stubs/nltk/tag/stanford.pyi +++ /dev/null @@ -1,40 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import ( - config_java as config_java, -) -from nltk.internals import ( - find_file as find_file, -) -from nltk.internals import ( - find_jar as find_jar, -) -from nltk.internals import ( - java as java, -) -from nltk.tag.api import TaggerI as TaggerI - -class StanfordTagger(TaggerI): - java_options: Incomplete - def __init__( - self, - model_filename: Incomplete, - path_to_jar: Incomplete | None = None, - encoding: str = "utf8", - verbose: bool = False, - java_options: str = "-mx1000m", - ) -> None: ... - def tag(self, tokens: Incomplete) -> Incomplete: ... - def tag_sents(self, sentences: Incomplete) -> Incomplete: ... - def parse_output( - self, text: Incomplete, sentences: Incomplete | None = None - ) -> Incomplete: ... - -class StanfordPOSTagger(StanfordTagger): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - -class StanfordNERTagger(StanfordTagger): - def __init__(self, *args: Incomplete, **kwargs: Incomplete) -> None: ... - def parse_output( - self, text: Incomplete, sentences: Incomplete - ) -> Incomplete: ... diff --git a/stubs/nltk/tag/tnt.pyi b/stubs/nltk/tag/tnt.pyi deleted file mode 100644 index 106cce8..0000000 --- a/stubs/nltk/tag/tnt.pyi +++ /dev/null @@ -1,28 +0,0 @@ -from _typeshed import Incomplete - -from nltk.probability import ( - ConditionalFreqDist as ConditionalFreqDist, -) -from nltk.probability import ( - FreqDist as FreqDist, -) -from nltk.tag.api import TaggerI as TaggerI - -class TnT(TaggerI): - unknown: int - known: int - def __init__( - self, - unk: Incomplete | None = None, - Trained: bool = False, - N: int = 1000, - C: bool = False, - ) -> None: ... - def train(self, data: Incomplete) -> None: ... - def tagdata(self, data: Incomplete) -> Incomplete: ... - def tag(self, data: Incomplete) -> Incomplete: ... - -def basic_sent_chop(data: Incomplete, raw: bool = True) -> Incomplete: ... -def demo() -> None: ... -def demo2() -> None: ... -def demo3() -> None: ... diff --git a/stubs/nltk/tag/util.pyi b/stubs/nltk/tag/util.pyi deleted file mode 100644 index a9aafd9..0000000 --- a/stubs/nltk/tag/util.pyi +++ /dev/null @@ -1,5 +0,0 @@ -from _typeshed import Incomplete - -def str2tuple(s: Incomplete, sep: str = "/") -> Incomplete: ... -def tuple2str(tagged_token: Incomplete, sep: str = "/") -> Incomplete: ... -def untag(tagged_sentence: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tbl/__init__.pyi b/stubs/nltk/tbl/__init__.pyi deleted file mode 100644 index 8990b7d..0000000 --- a/stubs/nltk/tbl/__init__.pyi +++ /dev/null @@ -1,4 +0,0 @@ -from nltk.tbl.erroranalysis import error_list as error_list -from nltk.tbl.feature import Feature as Feature -from nltk.tbl.rule import Rule as Rule -from nltk.tbl.template import Template as Template diff --git a/stubs/nltk/tbl/api.pyi b/stubs/nltk/tbl/api.pyi deleted file mode 100644 index e69de29..0000000 diff --git a/stubs/nltk/tbl/demo.pyi b/stubs/nltk/tbl/demo.pyi deleted file mode 100644 index bf715ab..0000000 --- a/stubs/nltk/tbl/demo.pyi +++ /dev/null @@ -1,55 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus import treebank as treebank -from nltk.tag import ( - BrillTaggerTrainer as BrillTaggerTrainer, -) -from nltk.tag import ( - RegexpTagger as RegexpTagger, -) -from nltk.tag import ( - UnigramTagger as UnigramTagger, -) -from nltk.tag.brill import Pos as Pos -from nltk.tag.brill import Word as Word -from nltk.tbl import Template as Template -from nltk.tbl import error_list as error_list - -def demo() -> None: ... -def demo_repr_rule_format() -> None: ... -def demo_str_rule_format() -> None: ... -def demo_verbose_rule_format() -> None: ... -def demo_multiposition_feature() -> None: ... -def demo_multifeature_template() -> None: ... -def demo_template_statistics() -> None: ... -def demo_generated_templates() -> None: ... -def demo_learning_curve() -> None: ... -def demo_error_analysis() -> None: ... -def demo_serialize_tagger() -> None: ... -def demo_high_accuracy_rules() -> None: ... -def postag( - templates: Incomplete | None = None, - tagged_data: Incomplete | None = None, - num_sents: int = 1000, - max_rules: int = 300, - min_score: int = 3, - min_acc: Incomplete | None = None, - train: float = 0.8, - trace: int = 3, - randomize: bool = False, - ruleformat: str = "str", - incremental_stats: bool = False, - template_stats: bool = False, - error_output: Incomplete | None = None, - serialize_output: Incomplete | None = None, - learning_curve_output: Incomplete | None = None, - learning_curve_take: int = 300, - baseline_backoff_tagger: Incomplete | None = None, - separate_baseline_data: bool = False, - cache_baseline_tagger: Incomplete | None = None, -) -> None: ... - -NN_CD_TAGGER: Incomplete -REGEXP_TAGGER: Incomplete - -def corpus_size(seqs: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tbl/erroranalysis.pyi b/stubs/nltk/tbl/erroranalysis.pyi deleted file mode 100644 index 8c8ba19..0000000 --- a/stubs/nltk/tbl/erroranalysis.pyi +++ /dev/null @@ -1,5 +0,0 @@ -from _typeshed import Incomplete - -def error_list( - train_sents: Incomplete, test_sents: Incomplete -) -> Incomplete: ... diff --git a/stubs/nltk/tbl/feature.pyi b/stubs/nltk/tbl/feature.pyi deleted file mode 100644 index 0d9da72..0000000 --- a/stubs/nltk/tbl/feature.pyi +++ /dev/null @@ -1,34 +0,0 @@ -from abc import ABCMeta, abstractmethod - -from _typeshed import Incomplete - -class Feature(metaclass=ABCMeta): - json_tag: str - PROPERTY_NAME: Incomplete - positions: Incomplete - def __init__( - self, positions: Incomplete, end: Incomplete | None = None - ) -> None: ... - def encode_json_obj(self) -> Incomplete: ... - @classmethod - def decode_json_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... - @classmethod - def expand( - cls: Incomplete, - starts: Incomplete, - winlens: Incomplete, - excludezero: bool = False, - ) -> Incomplete: ... - def issuperset(self, other: Incomplete) -> Incomplete: ... - def intersects(self, other: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __gt__(self, other: Incomplete) -> Incomplete: ... - def __ge__(self, other: Incomplete) -> Incomplete: ... - def __le__(self, other: Incomplete) -> Incomplete: ... - @staticmethod - @abstractmethod - def extract_property( - tokens: Incomplete, index: Incomplete - ) -> Incomplete: ... diff --git a/stubs/nltk/tbl/rule.pyi b/stubs/nltk/tbl/rule.pyi deleted file mode 100644 index b436678..0000000 --- a/stubs/nltk/tbl/rule.pyi +++ /dev/null @@ -1,39 +0,0 @@ -from abc import ABCMeta, abstractmethod - -from _typeshed import Incomplete - -from nltk import jsontags as jsontags - -class TagRule(metaclass=ABCMeta): - original_tag: Incomplete - replacement_tag: Incomplete - def __init__( - self, original_tag: Incomplete, replacement_tag: Incomplete - ) -> None: ... - def apply( - self, tokens: Incomplete, positions: Incomplete | None = None - ) -> Incomplete: ... - @abstractmethod - def applies(self, tokens: Incomplete, index: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - -class Rule(TagRule): - json_tag: str - templateid: Incomplete - def __init__( - self, - templateid: Incomplete, - original_tag: Incomplete, - replacement_tag: Incomplete, - conditions: Incomplete, - ) -> None: ... - def encode_json_obj(self) -> Incomplete: ... - @classmethod - def decode_json_obj(cls: Incomplete, obj: Incomplete) -> Incomplete: ... - def applies(self, tokens: Incomplete, index: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __ne__(self, other: Incomplete) -> Incomplete: ... - def __hash__(self) -> Incomplete: ... - def format(self, fmt: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tbl/template.pyi b/stubs/nltk/tbl/template.pyi deleted file mode 100644 index 189f7c8..0000000 --- a/stubs/nltk/tbl/template.pyi +++ /dev/null @@ -1,35 +0,0 @@ -from abc import ABCMeta, abstractmethod -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.tbl.feature import Feature as Feature -from nltk.tbl.rule import Rule as Rule - -class BrillTemplateI(metaclass=ABCMeta): - @abstractmethod - def applicable_rules( - self, tokens: Incomplete, i: Incomplete, correctTag: Incomplete - ) -> Incomplete: ... - @abstractmethod - def get_neighborhood( - self, token: Incomplete, index: Incomplete - ) -> Incomplete: ... - -class Template(BrillTemplateI): - ALLTEMPLATES: Incomplete - id: Incomplete - def __init__(self, *features: Incomplete) -> None: ... - def applicable_rules( - self, tokens: Incomplete, index: Incomplete, correct_tag: Incomplete - ) -> Incomplete: ... - def get_neighborhood( - self, tokens: Incomplete, index: Incomplete - ) -> Incomplete: ... - @classmethod - def expand( - cls: Incomplete, - featurelists: Incomplete, - combinations: Incomplete | None = None, - skipintersecting: bool = True, - ) -> Generator[Incomplete, None, Incomplete]: ... diff --git a/stubs/nltk/text.pyi b/stubs/nltk/text.pyi deleted file mode 100644 index 4112602..0000000 --- a/stubs/nltk/text.pyi +++ /dev/null @@ -1,92 +0,0 @@ -from typing import NamedTuple - -from _typeshed import Incomplete - -__all__ = [ - "ContextIndex", - "ConcordanceIndex", - "TokenSearcher", - "Text", - "TextCollection", -] - -class ConcordanceLine(NamedTuple): - left: Incomplete - query: Incomplete - right: Incomplete - offset: Incomplete - left_print: Incomplete - right_print: Incomplete - line: Incomplete - -class ContextIndex: - def __init__( - self, - tokens: Incomplete, - context_func: Incomplete | None = None, - filter: Incomplete | None = None, - key: Incomplete = ..., - ) -> None: ... - def tokens(self) -> Incomplete: ... - def word_similarity_dict(self, word: Incomplete) -> Incomplete: ... - def similar_words(self, word: Incomplete, n: int = 20) -> Incomplete: ... - def common_contexts( - self, words: Incomplete, fail_on_unknown: bool = False - ) -> Incomplete: ... - -class ConcordanceIndex: - def __init__(self, tokens: Incomplete, key: Incomplete = ...) -> None: ... - def tokens(self) -> Incomplete: ... - def offsets(self, word: Incomplete) -> Incomplete: ... - def find_concordance( - self, word: Incomplete, width: int = 80 - ) -> Incomplete: ... - def print_concordance( - self, word: Incomplete, width: int = 80, lines: int = 25 - ) -> None: ... - -class TokenSearcher: - def __init__(self, tokens: Incomplete) -> None: ... - def findall(self, regexp: Incomplete) -> Incomplete: ... - -class Text: - tokens: Incomplete - name: Incomplete - def __init__( - self, tokens: Incomplete, name: Incomplete | None = None - ) -> None: ... - def __getitem__(self, i: Incomplete) -> Incomplete: ... - def __len__(self) -> int: ... - def concordance( - self, word: Incomplete, width: int = 79, lines: int = 25 - ) -> Incomplete: ... - def concordance_list( - self, word: Incomplete, width: int = 79, lines: int = 25 - ) -> Incomplete: ... - def collocation_list( - self, num: int = 20, window_size: int = 2 - ) -> Incomplete: ... - def collocations(self, num: int = 20, window_size: int = 2) -> None: ... - def count(self, word: Incomplete) -> Incomplete: ... - def index(self, word: Incomplete) -> Incomplete: ... - def readability(self, method: Incomplete) -> None: ... - def similar(self, word: Incomplete, num: int = 20) -> Incomplete: ... - def common_contexts( - self, words: Incomplete, num: int = 20 - ) -> Incomplete: ... - def dispersion_plot(self, words: Incomplete) -> None: ... - def generate( - self, - length: int = 100, - text_seed: Incomplete | None = None, - random_seed: int = 42, - ) -> Incomplete: ... - def plot(self, *args: Incomplete) -> Incomplete: ... - def vocab(self) -> Incomplete: ... - def findall(self, regexp: Incomplete) -> None: ... - -class TextCollection(Text): - def __init__(self, source: Incomplete) -> None: ... - def tf(self, term: Incomplete, text: Incomplete) -> Incomplete: ... - def idf(self, term: Incomplete) -> Incomplete: ... - def tf_idf(self, term: Incomplete, text: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tgrep.pyi b/stubs/nltk/tgrep.pyi deleted file mode 100644 index 5232cbf..0000000 --- a/stubs/nltk/tgrep.pyi +++ /dev/null @@ -1,17 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -class TgrepException(Exception): ... - -def ancestors(node: Incomplete) -> Incomplete: ... -def unique_ancestors(node: Incomplete) -> Incomplete: ... -def tgrep_tokenize(tgrep_string: Incomplete) -> Incomplete: ... -def tgrep_compile(tgrep_string: Incomplete) -> Incomplete: ... -def treepositions_no_leaves(tree: Incomplete) -> Incomplete: ... -def tgrep_positions( - pattern: Incomplete, trees: Incomplete, search_leaves: bool = True -) -> Generator[Incomplete, None, None]: ... -def tgrep_nodes( - pattern: Incomplete, trees: Incomplete, search_leaves: bool = True -) -> Generator[Incomplete, None, None]: ... diff --git a/stubs/nltk/tokenize/__init__.pyi b/stubs/nltk/tokenize/__init__.pyi deleted file mode 100644 index 4085319..0000000 --- a/stubs/nltk/tokenize/__init__.pyi +++ /dev/null @@ -1,87 +0,0 @@ -from _typeshed import Incomplete - -from nltk.data import load as load -from nltk.tokenize.casual import ( - TweetTokenizer as TweetTokenizer, -) -from nltk.tokenize.casual import ( - casual_tokenize as casual_tokenize, -) -from nltk.tokenize.destructive import NLTKWordTokenizer as NLTKWordTokenizer -from nltk.tokenize.legality_principle import ( - LegalitySyllableTokenizer as LegalitySyllableTokenizer, -) -from nltk.tokenize.mwe import MWETokenizer as MWETokenizer -from nltk.tokenize.punkt import ( - PunktSentenceTokenizer as PunktSentenceTokenizer, -) -from nltk.tokenize.punkt import ( - PunktTokenizer as PunktTokenizer, -) -from nltk.tokenize.regexp import ( - BlanklineTokenizer as BlanklineTokenizer, -) -from nltk.tokenize.regexp import ( - RegexpTokenizer as RegexpTokenizer, -) -from nltk.tokenize.regexp import ( - WhitespaceTokenizer as WhitespaceTokenizer, -) -from nltk.tokenize.regexp import ( - WordPunctTokenizer as WordPunctTokenizer, -) -from nltk.tokenize.regexp import ( - blankline_tokenize as blankline_tokenize, -) -from nltk.tokenize.regexp import ( - regexp_tokenize as regexp_tokenize, -) -from nltk.tokenize.regexp import ( - wordpunct_tokenize as wordpunct_tokenize, -) -from nltk.tokenize.repp import ReppTokenizer as ReppTokenizer -from nltk.tokenize.sexpr import ( - SExprTokenizer as SExprTokenizer, -) -from nltk.tokenize.sexpr import ( - sexpr_tokenize as sexpr_tokenize, -) -from nltk.tokenize.simple import ( - LineTokenizer as LineTokenizer, -) -from nltk.tokenize.simple import ( - SpaceTokenizer as SpaceTokenizer, -) -from nltk.tokenize.simple import ( - TabTokenizer as TabTokenizer, -) -from nltk.tokenize.simple import ( - line_tokenize as line_tokenize, -) -from nltk.tokenize.sonority_sequencing import ( - SyllableTokenizer as SyllableTokenizer, -) -from nltk.tokenize.stanford_segmenter import ( - StanfordSegmenter as StanfordSegmenter, -) -from nltk.tokenize.texttiling import TextTilingTokenizer as TextTilingTokenizer -from nltk.tokenize.toktok import ToktokTokenizer as ToktokTokenizer -from nltk.tokenize.treebank import ( - TreebankWordDetokenizer as TreebankWordDetokenizer, -) -from nltk.tokenize.treebank import ( - TreebankWordTokenizer as TreebankWordTokenizer, -) -from nltk.tokenize.util import ( - regexp_span_tokenize as regexp_span_tokenize, -) -from nltk.tokenize.util import ( - string_span_tokenize as string_span_tokenize, -) - -def sent_tokenize( - text: Incomplete, language: str = "english" -) -> Incomplete: ... -def word_tokenize( - text: Incomplete, language: str = "english", preserve_line: bool = False -) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/api.pyi b/stubs/nltk/tokenize/api.pyi deleted file mode 100644 index d696a01..0000000 --- a/stubs/nltk/tokenize/api.pyi +++ /dev/null @@ -1,23 +0,0 @@ -from abc import ABC, abstractmethod -from collections.abc import Generator -from typing import Iterator - -from _typeshed import Incomplete - -from nltk.internals import overridden as overridden -from nltk.tokenize.util import string_span_tokenize as string_span_tokenize - -class TokenizerI(ABC): - @abstractmethod - def tokenize(self, s: str) -> list[str]: ... - def span_tokenize(self, s: str) -> Iterator[tuple[int, int]]: ... - def tokenize_sents(self, strings: list[str]) -> list[list[str]]: ... - def span_tokenize_sents( - self, strings: list[str] - ) -> Iterator[list[tuple[int, int]]]: ... - -class StringTokenizer(TokenizerI): - def tokenize(self, s: Incomplete) -> Incomplete: ... - def span_tokenize( - self, s: Incomplete - ) -> Generator[Incomplete, Incomplete, None]: ... diff --git a/stubs/nltk/tokenize/casual.pyi b/stubs/nltk/tokenize/casual.pyi deleted file mode 100644 index 47e9fc6..0000000 --- a/stubs/nltk/tokenize/casual.pyi +++ /dev/null @@ -1,43 +0,0 @@ -import regex -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI - -EMOTICONS: str -URLS: str -FLAGS: str -PHONE_REGEX: str -REGEXPS: Incomplete -REGEXPS_PHONE: Incomplete -HANG_RE: Incomplete -EMOTICON_RE: Incomplete -ENT_RE: Incomplete -HANDLES_RE: Incomplete - -class TweetTokenizer(TokenizerI): - preserve_case: Incomplete - reduce_len: Incomplete - strip_handles: Incomplete - match_phone_numbers: Incomplete - def __init__( - self, - preserve_case: bool = True, - reduce_len: bool = False, - strip_handles: bool = False, - match_phone_numbers: bool = True, - ) -> None: ... - def tokenize(self, text: str) -> list[str]: ... - @property - def WORD_RE(self) -> regex.Pattern: ... - @property - def PHONE_WORD_RE(self) -> regex.Pattern: ... - -def reduce_lengthening(text: Incomplete) -> Incomplete: ... -def remove_handles(text: Incomplete) -> Incomplete: ... -def casual_tokenize( - text: Incomplete, - preserve_case: bool = True, - reduce_len: bool = False, - strip_handles: bool = False, - match_phone_numbers: bool = True, -) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/destructive.pyi b/stubs/nltk/tokenize/destructive.pyi deleted file mode 100644 index 97ce4bd..0000000 --- a/stubs/nltk/tokenize/destructive.pyi +++ /dev/null @@ -1,28 +0,0 @@ -from typing import Iterator - -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI -from nltk.tokenize.util import align_tokens as align_tokens - -class MacIntyreContractions: - CONTRACTIONS2: Incomplete - CONTRACTIONS3: Incomplete - CONTRACTIONS4: Incomplete - -class NLTKWordTokenizer(TokenizerI): - STARTING_QUOTES: Incomplete - ENDING_QUOTES: Incomplete - PUNCTUATION: Incomplete - PARENS_BRACKETS: Incomplete - CONVERT_PARENTHESES: Incomplete - DOUBLE_DASHES: Incomplete - CONTRACTIONS2: Incomplete - CONTRACTIONS3: Incomplete - def tokenize( - self, - text: str, - convert_parentheses: bool = False, - return_str: bool = False, - ) -> list[str]: ... - def span_tokenize(self, text: str) -> Iterator[tuple[int, int]]: ... diff --git a/stubs/nltk/tokenize/legality_principle.pyi b/stubs/nltk/tokenize/legality_principle.pyi deleted file mode 100644 index 02916fd..0000000 --- a/stubs/nltk/tokenize/legality_principle.pyi +++ /dev/null @@ -1,17 +0,0 @@ -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI - -class LegalitySyllableTokenizer(TokenizerI): - legal_frequency_threshold: Incomplete - vowels: Incomplete - legal_onsets: Incomplete - def __init__( - self, - tokenized_source_text: Incomplete, - vowels: str = "aeiouy", - legal_frequency_threshold: float = 0.001, - ) -> None: ... - def find_legal_onsets(self, words: Incomplete) -> Incomplete: ... - def onset(self, word: Incomplete) -> Incomplete: ... - def tokenize(self, token: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/mwe.pyi b/stubs/nltk/tokenize/mwe.pyi deleted file mode 100644 index d340552..0000000 --- a/stubs/nltk/tokenize/mwe.pyi +++ /dev/null @@ -1,11 +0,0 @@ -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI -from nltk.util import Trie as Trie - -class MWETokenizer(TokenizerI): - def __init__( - self, mwes: Incomplete | None = None, separator: str = "_" - ) -> None: ... - def add_mwe(self, mwe: Incomplete) -> None: ... - def tokenize(self, text: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/nist.pyi b/stubs/nltk/tokenize/nist.pyi deleted file mode 100644 index d432d79..0000000 --- a/stubs/nltk/tokenize/nist.pyi +++ /dev/null @@ -1,40 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus import perluniprops as perluniprops -from nltk.tokenize.api import TokenizerI as TokenizerI -from nltk.tokenize.util import xml_unescape as xml_unescape - -class NISTTokenizer(TokenizerI): - STRIP_SKIP: Incomplete - STRIP_EOL_HYPHEN: Incomplete - PUNCT: Incomplete - PERIOD_COMMA_PRECEED: Incomplete - PERIOD_COMMA_FOLLOW: Incomplete - DASH_PRECEED_DIGIT: Incomplete - LANG_DEPENDENT_REGEXES: Incomplete - pup_number: Incomplete - pup_punct: Incomplete - pup_symbol: Incomplete - number_regex: Incomplete - punct_regex: Incomplete - symbol_regex: Incomplete - NONASCII: Incomplete - PUNCT_1: Incomplete - PUNCT_2: Incomplete - SYMBOLS: Incomplete - INTERNATIONAL_REGEXES: Incomplete - def lang_independent_sub(self, text: Incomplete) -> Incomplete: ... - def tokenize( - self, - text: Incomplete, - lowercase: bool = False, - western_lang: bool = True, - return_str: bool = False, - ) -> Incomplete: ... - def international_tokenize( - self, - text: Incomplete, - lowercase: bool = False, - split_non_ascii: bool = True, - return_str: bool = False, - ) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/punkt.pyi b/stubs/nltk/tokenize/punkt.pyi deleted file mode 100644 index 01668d0..0000000 --- a/stubs/nltk/tokenize/punkt.pyi +++ /dev/null @@ -1,170 +0,0 @@ -from re import Pattern -from typing import ( - Any, - Iterator, - List, - Set, - Tuple, - Type, - Union, -) - -from _typeshed import Incomplete - -from nltk.probability import FreqDist as FreqDist -from nltk.tokenize.api import TokenizerI as TokenizerI - -REASON_DEFAULT_DECISION: str -REASON_KNOWN_COLLOCATION: str -REASON_ABBR_WITH_ORTHOGRAPHIC_HEURISTIC: str -REASON_ABBR_WITH_SENTENCE_STARTER: str -REASON_INITIAL_WITH_ORTHOGRAPHIC_HEURISTIC: str -REASON_NUMBER_WITH_ORTHOGRAPHIC_HEURISTIC: str -REASON_INITIAL_WITH_SPECIAL_ORTHOGRAPHIC_HEURISTIC: str - -def _pair_iter(iterator: Union[str, List[PunktToken]]) -> Iterator[Any]: ... - -class PunktBaseClass: - def __init__( - self, - lang_vars: None = ..., - token_cls: Type[PunktToken] = ..., - params: None = ..., - ) -> None: ... - def _annotate_first_pass( - self, tokens: Iterator[PunktToken] - ) -> Iterator[PunktToken]: ... - def _first_pass_annotation(self, aug_tok: PunktToken) -> None: ... - def _tokenize_words(self, plaintext: str) -> Iterator[PunktToken]: ... - -class PunktLanguageVars: - @property - def _re_non_word_chars(self) -> str: ... - @property - def _re_sent_end_chars(self) -> str: ... - def _word_tokenizer_re(self) -> Pattern: ... - def period_context_re(self) -> Pattern: ... - def word_tokenize(self, s: str) -> List[str]: ... - -class PunktParameters: - def __init__(self) -> None: ... - def _debug_ortho_context(self, typ: str) -> Incomplete: ... - def add_ortho_context(self, typ: str, flag: int) -> Incomplete: ... - def clear_collocations(self) -> Incomplete: ... - def clear_sent_starters(self) -> Incomplete: ... - -class PunktToken: - def __init__(self, tok: str, **params: Incomplete) -> None: ... - def _get_type(self, tok: str) -> str: ... - @property - def first_case(self) -> Incomplete: ... - @property - def first_lower(self) -> bool: ... - @property - def first_upper(self) -> bool: ... - @property - def is_ellipsis(self) -> None: ... - @property - def is_initial(self) -> None: ... - @property - def is_number(self) -> bool: ... - @property - def type_no_period(self) -> str: ... - @property - def type_no_sentperiod(self) -> str: ... - -class PunktTrainer: - def __init__( - self, - train_text: Incomplete | None = None, - verbose: bool = False, - lang_vars: Incomplete | None = None, - token_cls: Type[PunktToken] = ..., - ) -> None: ... - @staticmethod - def _dunning_log_likelihood( - count_a: int, count_b: int, count_ab: int, N: int - ) -> float: ... - def _find_collocations(self) -> Incomplete: ... - def _find_sent_starters(self) -> Incomplete: ... - def _get_orthography_data( - self, tokens: List[PunktToken] - ) -> Incomplete: ... - def _get_sentbreak_count(self, tokens: List[PunktToken]) -> int: ... - def _reclassify_abbrev_types( - self, types: Set[str] - ) -> Iterator[Tuple[str, float, bool]]: ... - def _train_tokens( - self, tokens: Iterator[Any], verbose: bool - ) -> Incomplete: ... - def _unique_types(self, tokens: List[PunktToken]) -> Set[str]: ... - def get_params(self) -> Incomplete: ... - ABBREV: float - IGNORE_ABBREV_PENALTY: bool - ABBREV_BACKOFF: int - COLLOCATION: float - SENT_STARTER: int - INCLUDE_ALL_COLLOCS: bool - INCLUDE_ABBREV_COLLOCS: bool - MIN_COLLOC_FREQ: int - def train( - self, text: Incomplete, verbose: bool = False, finalize: bool = True - ) -> None: ... - def train_tokens( - self, tokens: Incomplete, verbose: bool = False, finalize: bool = True - ) -> None: ... - def finalize_training(self, verbose: bool = False) -> None: ... - def freq_threshold( - self, - ortho_thresh: int = 2, - type_thresh: int = 2, - colloc_thres: int = 2, - sentstart_thresh: int = 2, - ) -> None: ... - def find_abbrev_types(self) -> None: ... - -class PunktSentenceTokenizer(PunktBaseClass, TokenizerI): - def __init__( - self, - train_text: Incomplete | None = None, - verbose: bool = False, - lang_vars: Incomplete | None = None, - token_cls: Incomplete = ..., - ) -> None: ... - def train( - self, train_text: Incomplete, verbose: bool = False - ) -> Incomplete: ... - def tokenize( - self, text: str, realign_boundaries: bool = True - ) -> list[str]: ... - def debug_decisions(self, text: str) -> Iterator[dict[str, Any]]: ... - def span_tokenize( - self, text: str, realign_boundaries: bool = True - ) -> Iterator[tuple[int, int]]: ... - def sentences_from_text( - self, text: str, realign_boundaries: bool = True - ) -> list[str]: ... - def text_contains_sentbreak(self, text: str) -> bool: ... - def sentences_from_text_legacy(self, text: str) -> Iterator[str]: ... - def sentences_from_tokens( - self, tokens: Iterator[PunktToken] - ) -> Iterator[PunktToken]: ... - def dump(self, tokens: Iterator[PunktToken]) -> None: ... - PUNCTUATION: Incomplete - -class PunktTokenizer(PunktSentenceTokenizer): - def __init__(self, lang: str = "english") -> None: ... - def load_lang(self, lang: str = "english") -> None: ... - def save_params(self) -> None: ... - -def load_punkt_params(lang_dir: Incomplete) -> Incomplete: ... -def save_punkt_params( - params: Incomplete, dir: str = "/tmp/punkt_tab" -) -> None: ... - -DEBUG_DECISION_FMT: str - -def format_debug_decision(d: Incomplete) -> Incomplete: ... -def demo( - text: Incomplete, tok_cls: Incomplete = ..., train_cls: Incomplete = ... -) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/regexp.pyi b/stubs/nltk/tokenize/regexp.pyi deleted file mode 100644 index 19cd127..0000000 --- a/stubs/nltk/tokenize/regexp.pyi +++ /dev/null @@ -1,40 +0,0 @@ -from collections.abc import Generator -from re import RegexFlag - -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI -from nltk.tokenize.util import regexp_span_tokenize as regexp_span_tokenize - -class RegexpTokenizer(TokenizerI): - def __init__( - self, - pattern: str, - gaps: bool = False, - discard_empty: bool = True, - flags: RegexFlag = ..., - ) -> None: ... - def tokenize(self, text: Incomplete) -> Incomplete: ... - def span_tokenize( - self, text: Incomplete - ) -> Generator[Incomplete, None, None]: ... - -class WhitespaceTokenizer(RegexpTokenizer): - def __init__(self) -> None: ... - -class BlanklineTokenizer(RegexpTokenizer): - def __init__(self) -> None: ... - -class WordPunctTokenizer(RegexpTokenizer): - def __init__(self) -> None: ... - -def regexp_tokenize( - text: Incomplete, - pattern: Incomplete, - gaps: bool = False, - discard_empty: bool = True, - flags: Incomplete = ..., -) -> Incomplete: ... - -blankline_tokenize: Incomplete -wordpunct_tokenize: Incomplete diff --git a/stubs/nltk/tokenize/repp.pyi b/stubs/nltk/tokenize/repp.pyi deleted file mode 100644 index 1f32d84..0000000 --- a/stubs/nltk/tokenize/repp.pyi +++ /dev/null @@ -1,27 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.data import ZipFilePathPointer as ZipFilePathPointer -from nltk.internals import find_dir as find_dir -from nltk.tokenize.api import TokenizerI as TokenizerI - -class ReppTokenizer(TokenizerI): - repp_dir: Incomplete - working_dir: Incomplete - encoding: Incomplete - def __init__( - self, repp_dir: Incomplete, encoding: str = "utf8" - ) -> None: ... - def tokenize(self, sentence: Incomplete) -> Incomplete: ... - def tokenize_sents( - self, sentences: Incomplete, keep_token_positions: bool = False - ) -> Generator[Incomplete, None, None]: ... - def generate_repp_command( - self, inputfilename: Incomplete - ) -> Incomplete: ... - @staticmethod - def parse_repp_outputs( - repp_output: Incomplete, - ) -> Generator[Incomplete, None, None]: ... - def find_repptokenizer(self, repp_dirname: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/sexpr.pyi b/stubs/nltk/tokenize/sexpr.pyi deleted file mode 100644 index 407f1fd..0000000 --- a/stubs/nltk/tokenize/sexpr.pyi +++ /dev/null @@ -1,9 +0,0 @@ -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI - -class SExprTokenizer(TokenizerI): - def __init__(self, parens: str = "()", strict: bool = True) -> None: ... - def tokenize(self, text: Incomplete) -> Incomplete: ... - -sexpr_tokenize: Incomplete diff --git a/stubs/nltk/tokenize/simple.pyi b/stubs/nltk/tokenize/simple.pyi deleted file mode 100644 index fc345ff..0000000 --- a/stubs/nltk/tokenize/simple.pyi +++ /dev/null @@ -1,34 +0,0 @@ -from typing import ( - Iterator, - List, - Tuple, -) - -from _typeshed import Incomplete - -from nltk.tokenize.api import ( - StringTokenizer as StringTokenizer, -) -from nltk.tokenize.api import ( - TokenizerI as TokenizerI, -) -from nltk.tokenize.util import ( - regexp_span_tokenize as regexp_span_tokenize, -) -from nltk.tokenize.util import ( - string_span_tokenize as string_span_tokenize, -) - -class CharTokenizer: - def span_tokenize(self, s: str) -> Iterator[Tuple[int, int]]: ... - def tokenize(self, s: str) -> List[str]: ... - -class LineTokenizer: - def __init__(self, blanklines: str = ...): ... - -class SpaceTokenizer(StringTokenizer): ... -class TabTokenizer(StringTokenizer): ... - -def line_tokenize( - text: Incomplete, blanklines: str = "discard" -) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/sonority_sequencing.pyi b/stubs/nltk/tokenize/sonority_sequencing.pyi deleted file mode 100644 index 2733b9f..0000000 --- a/stubs/nltk/tokenize/sonority_sequencing.pyi +++ /dev/null @@ -1,19 +0,0 @@ -from typing import ( - List, - Tuple, -) - -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI -from nltk.util import ngrams as ngrams - -class SyllableTokenizer: - vowels: Incomplete - phoneme_map: Incomplete - def __init__( - self, lang: str = "en", sonority_hierarchy: bool = ... - ) -> None: ... - def assign_values(self, token: str) -> List[Tuple[str, int]]: ... - def tokenize(self, token: str) -> List[str]: ... - def validate_syllables(self, syllable_list: List[str]) -> List[str]: ... diff --git a/stubs/nltk/tokenize/stanford.pyi b/stubs/nltk/tokenize/stanford.pyi deleted file mode 100644 index a546e72..0000000 --- a/stubs/nltk/tokenize/stanford.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import ( - config_java as config_java, -) -from nltk.internals import ( - find_jar as find_jar, -) -from nltk.internals import ( - java as java, -) -from nltk.parse.corenlp import CoreNLPParser as CoreNLPParser -from nltk.tokenize.api import TokenizerI as TokenizerI - -class StanfordTokenizer(TokenizerI): - java_options: Incomplete - def __init__( - self, - path_to_jar: Incomplete | None = None, - encoding: str = "utf8", - options: Incomplete | None = None, - verbose: bool = False, - java_options: str = "-mx1000m", - ) -> None: ... - def tokenize(self, s: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/stanford_segmenter.pyi b/stubs/nltk/tokenize/stanford_segmenter.pyi deleted file mode 100644 index 968bff2..0000000 --- a/stubs/nltk/tokenize/stanford_segmenter.pyi +++ /dev/null @@ -1,41 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import ( - config_java as config_java, -) -from nltk.internals import ( - find_dir as find_dir, -) -from nltk.internals import ( - find_file as find_file, -) -from nltk.internals import ( - find_jar as find_jar, -) -from nltk.internals import ( - java as java, -) -from nltk.tokenize.api import TokenizerI as TokenizerI - -class StanfordSegmenter: - java_options: Incomplete - def __init__( - self, - path_to_jar: None = ..., - path_to_slf4j: None = ..., - java_class: None = ..., - path_to_model: None = ..., - path_to_dict: None = ..., - path_to_sihan_corpora_dict: None = ..., - sihan_post_processing: str = ..., - keep_whitespaces: str = ..., - encoding: str = ..., - options: None = ..., - verbose: bool = ..., - java_options: str = ..., - ) -> None: ... - def default_config(self, lang: Incomplete) -> None: ... - def tokenize(self, s: Incomplete) -> None: ... - def segment_file(self, input_file_path: Incomplete) -> Incomplete: ... - def segment(self, tokens: Incomplete) -> Incomplete: ... - def segment_sents(self, sentences: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/texttiling.pyi b/stubs/nltk/tokenize/texttiling.pyi deleted file mode 100644 index 139eb49..0000000 --- a/stubs/nltk/tokenize/texttiling.pyi +++ /dev/null @@ -1,48 +0,0 @@ -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI - -BLOCK_COMPARISON: Incomplete -VOCABULARY_INTRODUCTION: Incomplete -LC: Incomplete -HC: Incomplete -DEFAULT_SMOOTHING: Incomplete - -class TextTilingTokenizer(TokenizerI): - def __init__( - self, - w: int = 20, - k: int = 10, - similarity_method: Incomplete = ..., - stopwords: Incomplete | None = None, - smoothing_method: Incomplete = ..., - smoothing_width: int = 2, - smoothing_rounds: int = 1, - cutoff_policy: Incomplete = ..., - demo_mode: bool = False, - ) -> None: ... - def tokenize(self, text: Incomplete) -> Incomplete: ... - -class TokenTableField: - def __init__( - self, - first_pos: Incomplete, - ts_occurences: Incomplete, - total_count: int = 1, - par_count: int = 1, - last_par: int = 0, - last_tok_seq: Incomplete | None = None, - ) -> None: ... - -class TokenSequence: - def __init__( - self, - index: Incomplete, - wrdindex_list: Incomplete, - original_length: Incomplete | None = None, - ) -> None: ... - -def smooth( - x: Incomplete, window_len: int = 11, window: str = "flat" -) -> Incomplete: ... -def demo(text: Incomplete | None = None) -> None: ... diff --git a/stubs/nltk/tokenize/toktok.pyi b/stubs/nltk/tokenize/toktok.pyi deleted file mode 100644 index 209ea3b..0000000 --- a/stubs/nltk/tokenize/toktok.pyi +++ /dev/null @@ -1,38 +0,0 @@ -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI - -class ToktokTokenizer(TokenizerI): - NON_BREAKING: Incomplete - FUNKY_PUNCT_1: Incomplete - FUNKY_PUNCT_2: Incomplete - EN_EM_DASHES: Incomplete - AMPERCENT: Incomplete - TAB: Incomplete - PIPE: Incomplete - COMMA_IN_NUM: Incomplete - PROB_SINGLE_QUOTES: Incomplete - STUPID_QUOTES_1: Incomplete - STUPID_QUOTES_2: Incomplete - FINAL_PERIOD_1: Incomplete - FINAL_PERIOD_2: Incomplete - MULTI_COMMAS: Incomplete - MULTI_DASHES: Incomplete - MULTI_DOTS: Incomplete - OPEN_PUNCT: Incomplete - CLOSE_PUNCT: Incomplete - CURRENCY_SYM: Incomplete - OPEN_PUNCT_RE: Incomplete - CLOSE_PUNCT_RE: Incomplete - CURRENCY_SYM_RE: Incomplete - URL_FOE_1: Incomplete - URL_FOE_2: Incomplete - URL_FOE_3: Incomplete - URL_FOE_4: Incomplete - LSTRIP: Incomplete - RSTRIP: Incomplete - ONE_SPACE: Incomplete - TOKTOK_REGEXES: Incomplete - def tokenize( - self, text: Incomplete, return_str: bool = False - ) -> Incomplete: ... diff --git a/stubs/nltk/tokenize/treebank.pyi b/stubs/nltk/tokenize/treebank.pyi deleted file mode 100644 index 0834ef9..0000000 --- a/stubs/nltk/tokenize/treebank.pyi +++ /dev/null @@ -1,38 +0,0 @@ -from typing import ( - Iterator, - List, - Tuple, -) - -from _typeshed import Incomplete - -from nltk.tokenize.api import TokenizerI as TokenizerI -from nltk.tokenize.destructive import ( - MacIntyreContractions as MacIntyreContractions, -) -from nltk.tokenize.util import align_tokens as align_tokens - -class TreebankWordTokenizer: - def span_tokenize(self, text: str) -> Iterator[Tuple[int, int]]: ... - def tokenize( - self, - text: str, - convert_parentheses: bool = ..., - return_str: bool = ..., - ) -> List[str]: ... - -class TreebankWordDetokenizer(TokenizerI): - CONTRACTIONS2: Incomplete - CONTRACTIONS3: Incomplete - ENDING_QUOTES: Incomplete - DOUBLE_DASHES: Incomplete - CONVERT_PARENTHESES: Incomplete - PARENS_BRACKETS: Incomplete - PUNCTUATION: Incomplete - STARTING_QUOTES: Incomplete - def tokenize( - self, tokens: list[str], convert_parentheses: bool = False - ) -> str: ... - def detokenize( - self, tokens: list[str], convert_parentheses: bool = False - ) -> str: ... diff --git a/stubs/nltk/tokenize/util.pyi b/stubs/nltk/tokenize/util.pyi deleted file mode 100644 index ce44fb1..0000000 --- a/stubs/nltk/tokenize/util.pyi +++ /dev/null @@ -1,35 +0,0 @@ -from collections.abc import Generator -from typing import ( - List, - Tuple, -) - -from _typeshed import Incomplete - -def align_tokens( - tokens: List[str], sentence: str -) -> List[Tuple[int, int]]: ... -def string_span_tokenize( - s: Incomplete, sep: Incomplete -) -> Generator[Incomplete, None, None]: ... -def regexp_span_tokenize( - s: Incomplete, regexp: Incomplete -) -> Generator[Incomplete, None, None]: ... -def spans_to_relative( - spans: Incomplete, -) -> Generator[Incomplete, None, None]: ... - -class CJKChars: - Hangul_Jamo: Incomplete - CJK_Radicals: Incomplete - Phags_Pa: Incomplete - Hangul_Syllables: Incomplete - CJK_Compatibility_Ideographs: Incomplete - CJK_Compatibility_Forms: Incomplete - Katakana_Hangul_Halfwidth: Incomplete - Supplementary_Ideographic_Plane: Incomplete - ranges: Incomplete - -def is_cjk(character: Incomplete) -> Incomplete: ... -def xml_escape(text: Incomplete) -> Incomplete: ... -def xml_unescape(text: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/toolbox.pyi b/stubs/nltk/toolbox.pyi deleted file mode 100644 index 34d4777..0000000 --- a/stubs/nltk/toolbox.pyi +++ /dev/null @@ -1,63 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete - -from nltk.data import PathPointer as PathPointer -from nltk.data import find as find - -class StandardFormat: - def __init__( - self, - filename: Incomplete | None = None, - encoding: Incomplete | None = None, - ) -> None: ... - def open(self, sfm_file: Incomplete) -> None: ... - def open_string(self, s: Incomplete) -> None: ... - line_num: int - def raw_fields(self) -> Generator[Incomplete, None, None]: ... - def fields( - self, - strip: bool = True, - unwrap: bool = True, - encoding: Incomplete | None = None, - errors: str = "strict", - unicode_fields: Incomplete | None = None, - ) -> Generator[Incomplete, None, None]: ... - def close(self) -> None: ... - -class ToolboxData(StandardFormat): - def parse( - self, grammar: Incomplete | None = None, **kwargs: Incomplete - ) -> Incomplete: ... - -def to_sfm_string( - tree: Incomplete, - encoding: Incomplete | None = None, - errors: str = "strict", - unicode_fields: Incomplete | None = None, -) -> Incomplete: ... - -class ToolboxSettings(StandardFormat): - def __init__(self) -> None: ... - def parse( - self, - encoding: Incomplete | None = None, - errors: str = "strict", - **kwargs: Incomplete, - ) -> Incomplete: ... - -def to_settings_string( - tree: Incomplete, - encoding: Incomplete | None = None, - errors: str = "strict", - unicode_fields: Incomplete | None = None, -) -> Incomplete: ... -def remove_blanks(elem: Incomplete) -> None: ... -def add_default_fields( - elem: Incomplete, default_fields: Incomplete -) -> None: ... -def sort_fields(elem: Incomplete, field_orders: Incomplete) -> None: ... -def add_blank_lines( - tree: Incomplete, blanks_before: Incomplete, blanks_between: Incomplete -) -> None: ... -def demo() -> None: ... diff --git a/stubs/nltk/translate/__init__.pyi b/stubs/nltk/translate/__init__.pyi deleted file mode 100644 index 574c338..0000000 --- a/stubs/nltk/translate/__init__.pyi +++ /dev/null @@ -1,20 +0,0 @@ -from nltk.translate.api import ( - AlignedSent as AlignedSent, -) -from nltk.translate.api import ( - Alignment as Alignment, -) -from nltk.translate.api import ( - PhraseTable as PhraseTable, -) -from nltk.translate.gale_church import trace as trace -from nltk.translate.gdfa import grow_diag_final_and as grow_diag_final_and -from nltk.translate.ibm1 import IBMModel1 as IBMModel1 -from nltk.translate.ibm2 import IBMModel2 as IBMModel2 -from nltk.translate.ibm3 import IBMModel3 as IBMModel3 -from nltk.translate.ibm4 import IBMModel4 as IBMModel4 -from nltk.translate.ibm5 import IBMModel5 as IBMModel5 -from nltk.translate.ibm_model import IBMModel as IBMModel -from nltk.translate.metrics import alignment_error_rate as alignment_error_rate -from nltk.translate.phrase_based import extract as extract -from nltk.translate.stack_decoder import StackDecoder as StackDecoder diff --git a/stubs/nltk/translate/api.pyi b/stubs/nltk/translate/api.pyi deleted file mode 100644 index 2392d83..0000000 --- a/stubs/nltk/translate/api.pyi +++ /dev/null @@ -1,49 +0,0 @@ -from typing import ( - Any, - List, - NamedTuple, - Tuple, - Type, - Union, -) - -from _typeshed import Incomplete - -PhraseTableEntry: NamedTuple - -def _check_alignment( - num_words: int, num_mots: int, alignment: Alignment -) -> Incomplete: ... - -class AlignedSent: - def __init__( - self, - words: List[Union[str, Any]], - mots: List[Union[str, Any]], - alignment: None = ..., - ) -> None: ... - def _set_alignment(self, alignment: Alignment) -> Incomplete: ... - @property - def mots(self) -> List[Union[str, Any]]: ... - @property - def words(self) -> List[Union[str, Any]]: ... - -class Alignment: - @staticmethod - def __new__( - cls: Type[Alignment], pairs: List[Union[Tuple[int, int], Any]] - ) -> Alignment: ... - -class PhraseTable: - def __contains__(self, src_phrase: Tuple[str, Ellipsis]) -> bool: ... # type: ignore[valid-type] - def __init__(self) -> None: ... - def add( - self, - src_phrase: Union[Tuple[str, ...]], - trg_phrase: Union[Tuple[str, ...]], - log_prob: float, - ) -> Incomplete: ... - def translations_for( - self, - src_phrase: Union[Tuple[str, ...]], - ) -> List[PhraseTableEntry]: ... # type: ignore[valid-type] diff --git a/stubs/nltk/translate/bleu_score.pyi b/stubs/nltk/translate/bleu_score.pyi deleted file mode 100644 index aa92c64..0000000 --- a/stubs/nltk/translate/bleu_score.pyi +++ /dev/null @@ -1,64 +0,0 @@ -from typing import ( - Any, - Callable, - List, - Optional, - Tuple, - Type, - Union, -) - -from _typeshed import Incomplete -from numpy import ndarray - -from nltk.util import ngrams as ngrams - -def brevity_penalty(closest_ref_len: int, hyp_len: int) -> float: ... -def closest_ref_length( - references: List[List[Union[str, Any]]], hyp_len: int -) -> int: ... -def corpus_bleu( - list_of_references: List[List[List[Union[str, Any]]]], - hypotheses: List[List[Union[str, Any]]], - weights: Any = ..., - smoothing_function: Optional[Callable] = ..., # type: ignore[type-arg] - auto_reweigh: bool = ..., -) -> Union[float, List[float], int]: ... -def modified_precision( - references: List[List[Union[str, Any]]], - hypothesis: List[Union[str, Any]], - n: int, -) -> Fraction: ... -def sentence_bleu( - references: List[List[Union[str, Any]]], - hypothesis: List[Union[str, Any]], - weights: Union[ - Tuple[float, float, float, float], - Tuple[float, float], - ndarray, - Tuple[float], - ] = ..., - smoothing_function: Optional[Callable] = ..., # type: ignore[type-arg] - auto_reweigh: bool = ..., -) -> Union[int, float]: ... - -class Fraction: - @staticmethod - def __new__( - cls: Type[Fraction], - numerator: int = ..., - denominator: Optional[int] = ..., - _normalize: bool = ..., - ) -> Fraction: ... - @property - def denominator(self) -> int: ... - @property - def numerator(self) -> int: ... - -class SmoothingFunction: - def __init__( - self, epsilon: float = ..., alpha: int = ..., k: int = ... - ) -> None: ... - def method0( - self, p_n: List[Fraction], *args: Incomplete, **kwargs: Incomplete - ) -> List[Union[Fraction, float]]: ... diff --git a/stubs/nltk/translate/chrf_score.pyi b/stubs/nltk/translate/chrf_score.pyi deleted file mode 100644 index 867588e..0000000 --- a/stubs/nltk/translate/chrf_score.pyi +++ /dev/null @@ -1,27 +0,0 @@ -from _typeshed import Incomplete - -from nltk.util import ngrams as ngrams - -def sentence_chrf( - reference: Incomplete, - hypothesis: Incomplete, - min_len: int = 1, - max_len: int = 6, - beta: float = 3.0, - ignore_whitespace: bool = True, -) -> Incomplete: ... -def chrf_precision_recall_fscore_support( - reference: Incomplete, - hypothesis: Incomplete, - n: Incomplete, - beta: float = 3.0, - epsilon: float = 1e-16, -) -> Incomplete: ... -def corpus_chrf( - references: Incomplete, - hypotheses: Incomplete, - min_len: int = 1, - max_len: int = 6, - beta: float = 3.0, - ignore_whitespace: bool = True, -) -> Incomplete: ... diff --git a/stubs/nltk/translate/gale_church.pyi b/stubs/nltk/translate/gale_church.pyi deleted file mode 100644 index 367d362..0000000 --- a/stubs/nltk/translate/gale_church.pyi +++ /dev/null @@ -1,44 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete -from scipy.stats import norm as norm # type: ignore[import-not-found] - -def erfcc(x: Incomplete) -> Incomplete: ... -def norm_cdf(x: Incomplete) -> Incomplete: ... - -LOG2: Incomplete - -class LanguageIndependent: - PRIORS: Incomplete - AVERAGE_CHARACTERS: int - VARIANCE_CHARACTERS: float - -def trace( - backlinks: Incomplete, - source_sents_lens: Incomplete, - target_sents_lens: Incomplete, -) -> Incomplete: ... -def align_log_prob( - i: Incomplete, - j: Incomplete, - source_sents: Incomplete, - target_sents: Incomplete, - alignment: Incomplete, - params: Incomplete, -) -> Incomplete: ... -def align_blocks( - source_sents_lens: Incomplete, - target_sents_lens: Incomplete, - params: Incomplete = ..., -) -> Incomplete: ... -def align_texts( - source_blocks: Incomplete, - target_blocks: Incomplete, - params: Incomplete = ..., -) -> Incomplete: ... -def split_at( - it: Incomplete, split_value: Incomplete -) -> Generator[Incomplete, None, None]: ... -def parse_token_stream( - stream: Incomplete, soft_delimiter: Incomplete, hard_delimiter: Incomplete -) -> Incomplete: ... diff --git a/stubs/nltk/translate/gdfa.pyi b/stubs/nltk/translate/gdfa.pyi deleted file mode 100644 index b1fad7d..0000000 --- a/stubs/nltk/translate/gdfa.pyi +++ /dev/null @@ -1,8 +0,0 @@ -from typing import ( - List, - Tuple, -) - -def grow_diag_final_and( - srclen: int, trglen: int, e2f: str, f2e: str -) -> List[Tuple[int, int]]: ... diff --git a/stubs/nltk/translate/gleu_score.pyi b/stubs/nltk/translate/gleu_score.pyi deleted file mode 100644 index 030326a..0000000 --- a/stubs/nltk/translate/gleu_score.pyi +++ /dev/null @@ -1,17 +0,0 @@ -from _typeshed import Incomplete - -from nltk.util import everygrams as everygrams -from nltk.util import ngrams as ngrams - -def sentence_gleu( - references: Incomplete, - hypothesis: Incomplete, - min_len: int = 1, - max_len: int = 4, -) -> Incomplete: ... -def corpus_gleu( - list_of_references: Incomplete, - hypotheses: Incomplete, - min_len: int = 1, - max_len: int = 4, -) -> Incomplete: ... diff --git a/stubs/nltk/translate/ibm1.pyi b/stubs/nltk/translate/ibm1.pyi deleted file mode 100644 index 9081d3b..0000000 --- a/stubs/nltk/translate/ibm1.pyi +++ /dev/null @@ -1,31 +0,0 @@ -from typing import ( - List, -) - -from _typeshed import Incomplete - -from nltk.translate import ( - AlignedSent as AlignedSent, -) -from nltk.translate import ( - Alignment as Alignment, -) -from nltk.translate import ( - IBMModel as IBMModel, -) -from nltk.translate.ibm_model import AlignmentInfo -from nltk.translate.ibm_model import Counts as Counts - -class IBMModel1: - def __init__( - self, - sentence_aligned_corpus: List[AlignedSent], - iterations: int, - probability_tables: None = ..., - ) -> None: ... - def align(self, sentence_pair: AlignedSent) -> Incomplete: ... - def align_all(self, parallel_corpus: List[AlignedSent]) -> Incomplete: ... - def prob_t_a_given_s(self, alignment_info: AlignmentInfo) -> float: ... - def set_uniform_probabilities( - self, sentence_aligned_corpus: List[AlignedSent] - ) -> Incomplete: ... diff --git a/stubs/nltk/translate/ibm2.pyi b/stubs/nltk/translate/ibm2.pyi deleted file mode 100644 index 682fa47..0000000 --- a/stubs/nltk/translate/ibm2.pyi +++ /dev/null @@ -1,50 +0,0 @@ -from typing import ( - List, -) - -from _typeshed import Incomplete - -from nltk.translate import ( - AlignedSent as AlignedSent, -) -from nltk.translate import ( - Alignment as Alignment, -) -from nltk.translate import ( - IBMModel as IBMModel, -) -from nltk.translate import ( - IBMModel1 as IBMModel1, -) -from nltk.translate.ibm_model import AlignmentInfo -from nltk.translate.ibm_model import Counts as Counts - -class IBMModel2: - def __init__( - self, - sentence_aligned_corpus: List[AlignedSent], - iterations: int, - probability_tables: None = ..., - ) -> None: ... - def align(self, sentence_pair: AlignedSent) -> Incomplete: ... - def align_all(self, parallel_corpus: List[AlignedSent]) -> Incomplete: ... - def prob_t_a_given_s(self, alignment_info: AlignmentInfo) -> float: ... - def set_uniform_probabilities( - self, sentence_aligned_corpus: List[AlignedSent] - ) -> Incomplete: ... - -class Model2Counts(Counts): - alignment: Incomplete - alignment_for_any_i: Incomplete - def __init__(self) -> None: ... - def update_lexical_translation( - self, count: Incomplete, s: Incomplete, t: Incomplete - ) -> None: ... - def update_alignment( - self, - count: Incomplete, - i: Incomplete, - j: Incomplete, - l: Incomplete, - m: Incomplete, - ) -> None: ... diff --git a/stubs/nltk/translate/ibm3.pyi b/stubs/nltk/translate/ibm3.pyi deleted file mode 100644 index 7145597..0000000 --- a/stubs/nltk/translate/ibm3.pyi +++ /dev/null @@ -1,79 +0,0 @@ -from typing import ( - DefaultDict, - Dict, - List, - Optional, - Union, -) - -from _typeshed import Incomplete - -from nltk.translate import ( - AlignedSent as AlignedSent, -) -from nltk.translate import ( - Alignment as Alignment, -) -from nltk.translate import ( - IBMModel as IBMModel, -) -from nltk.translate import ( - IBMModel2 as IBMModel2, -) -from nltk.translate.ibm_model import AlignmentInfo -from nltk.translate.ibm_model import Counts as Counts - -class IBMModel3: - def __init__( - self, - sentence_aligned_corpus: List[AlignedSent], - iterations: int, - probability_tables: Optional[ - Dict[ - str, - Optional[ - Union[ - float, - DefaultDict[ - str, - Union[ - DefaultDict[str, float], - DefaultDict[None, float], - ], - ], - DefaultDict[ - int, - DefaultDict[ - int, DefaultDict[int, DefaultDict[int, float]] - ], - ], - DefaultDict[ - int, - Union[ - DefaultDict[Optional[str], float], - DefaultDict[str, float], - ], - ], - ] - ], - ] - ] = ..., - ) -> None: ... - def prob_t_a_given_s(self, alignment_info: AlignmentInfo) -> float: ... - def reset_probabilities(self) -> Incomplete: ... - def set_uniform_probabilities( - self, sentence_aligned_corpus: List[AlignedSent] - ) -> Incomplete: ... - -class Model3Counts(Counts): - distortion: Incomplete - distortion_for_any_j: Incomplete - def __init__(self) -> None: ... - def update_distortion( - self, - count: Incomplete, - alignment_info: Incomplete, - j: Incomplete, - l: Incomplete, - m: Incomplete, - ) -> None: ... diff --git a/stubs/nltk/translate/ibm4.pyi b/stubs/nltk/translate/ibm4.pyi deleted file mode 100644 index ca32a19..0000000 --- a/stubs/nltk/translate/ibm4.pyi +++ /dev/null @@ -1,94 +0,0 @@ -from typing import ( - DefaultDict, - Dict, - List, - Optional, - Union, -) - -from _typeshed import Incomplete - -from nltk.translate import ( - AlignedSent as AlignedSent, -) -from nltk.translate import ( - Alignment as Alignment, -) -from nltk.translate import ( - IBMModel as IBMModel, -) -from nltk.translate import ( - IBMModel3 as IBMModel3, -) -from nltk.translate.ibm_model import AlignmentInfo -from nltk.translate.ibm_model import ( - Counts as Counts, -) -from nltk.translate.ibm_model import ( - longest_target_sentence_length as longest_target_sentence_length, -) - -class IBMModel4: - def __init__( - self, - sentence_aligned_corpus: List[AlignedSent], - iterations: int, - source_word_classes: Optional[Dict[str, int]], - target_word_classes: Optional[Dict[str, int]], - probability_tables: Optional[ - Dict[ - str, - Optional[ - Union[ - float, - DefaultDict[ - str, - Union[ - DefaultDict[str, float], - DefaultDict[None, float], - ], - ], - DefaultDict[ - int, - Union[ - DefaultDict[None, DefaultDict[int, float]], - DefaultDict[int, DefaultDict[int, float]], - ], - ], - DefaultDict[int, DefaultDict[int, float]], - DefaultDict[ - int, - Union[ - DefaultDict[Optional[str], float], - DefaultDict[str, float], - ], - ], - ] - ], - ] - ] = ..., - ) -> None: ... - @staticmethod - def model4_prob_t_a_given_s( - alignment_info: AlignmentInfo, ibm_model: IBMModel4 - ) -> float: ... - def prob_t_a_given_s(self, alignment_info: AlignmentInfo) -> float: ... - def reset_probabilities(self) -> Incomplete: ... - def set_uniform_probabilities( - self, sentence_aligned_corpus: List[AlignedSent] - ) -> Incomplete: ... - -class Model4Counts(Counts): - head_distortion: Incomplete - head_distortion_for_any_dj: Incomplete - non_head_distortion: Incomplete - non_head_distortion_for_any_dj: Incomplete - def __init__(self) -> None: ... - def update_distortion( - self, - count: Incomplete, - alignment_info: Incomplete, - j: Incomplete, - src_classes: Incomplete, - trg_classes: Incomplete, - ) -> None: ... diff --git a/stubs/nltk/translate/ibm5.pyi b/stubs/nltk/translate/ibm5.pyi deleted file mode 100644 index c212dc1..0000000 --- a/stubs/nltk/translate/ibm5.pyi +++ /dev/null @@ -1,95 +0,0 @@ -from typing import ( - DefaultDict, - Dict, - List, - Optional, - Set, - Union, -) - -from _typeshed import Incomplete - -from nltk.translate import ( - AlignedSent as AlignedSent, -) -from nltk.translate import ( - Alignment as Alignment, -) -from nltk.translate import ( - IBMModel as IBMModel, -) -from nltk.translate import ( - IBMModel4 as IBMModel4, -) -from nltk.translate.ibm_model import AlignmentInfo -from nltk.translate.ibm_model import ( - Counts as Counts, -) -from nltk.translate.ibm_model import ( - longest_target_sentence_length as longest_target_sentence_length, -) - -class IBMModel5: - def __init__( - self, - sentence_aligned_corpus: List[AlignedSent], - iterations: int, - source_word_classes: Optional[Dict[str, int]], - target_word_classes: Optional[Dict[str, int]], - probability_tables: Optional[ - Dict[ - str, - Optional[ - Union[ - float, - DefaultDict[ - str, - Union[ - DefaultDict[str, float], - DefaultDict[None, float], - ], - ], - DefaultDict[ - int, - Union[ - DefaultDict[Optional[str], float], - DefaultDict[str, float], - ], - ], - DefaultDict[ - int, DefaultDict[int, DefaultDict[int, float]] - ], - ] - ], - ] - ] = ..., - ) -> None: ... - def prob_t_a_given_s(self, alignment_info: AlignmentInfo) -> float: ... - def prune( - self, alignment_infos: List[AlignmentInfo] - ) -> Set[AlignmentInfo]: ... - def reset_probabilities(self) -> Incomplete: ... - def set_uniform_probabilities( - self, sentence_aligned_corpus: List[AlignedSent] - ) -> Incomplete: ... - -class Slots: - def __init__(self, target_sentence_length: int) -> None: ... - def __len__(self) -> int: ... - def occupy(self, position: int) -> Incomplete: ... - def vacancies_at(self, position: int) -> int: ... - -class Model5Counts(Counts): - head_vacancy: Incomplete - head_vacancy_for_any_dv: Incomplete - non_head_vacancy: Incomplete - non_head_vacancy_for_any_dv: Incomplete - def __init__(self) -> None: ... - def update_vacancy( - self, - count: Incomplete, - alignment_info: Incomplete, - i: Incomplete, - trg_classes: Incomplete, - slots: Incomplete, - ) -> None: ... diff --git a/stubs/nltk/translate/ibm_model.pyi b/stubs/nltk/translate/ibm_model.pyi deleted file mode 100644 index 47fc184..0000000 --- a/stubs/nltk/translate/ibm_model.pyi +++ /dev/null @@ -1,91 +0,0 @@ -from typing import ( - Any, - List, - Optional, - Set, - Tuple, - Union, -) - -from _typeshed import Incomplete - -from nltk.translate.api import AlignedSent - -def longest_target_sentence_length( - sentence_aligned_corpus: List[AlignedSent], -) -> int: ... - -class AlignmentInfo: - def __eq__(self, other: AlignmentInfo) -> bool: ... - def __hash__(self) -> int: ... - def __init__( - self, - alignment: Union[ - Tuple[int, int, int], - Tuple[int, int], - Tuple[int, int, int, int], - Tuple[int], - Tuple[int, int, int, int, int, int, int], - ], - src_sentence: Optional[ - Union[ - Tuple[None, str, str, str, str], - Tuple[None, str, str, str], - List[Optional[str]], - Tuple[None], - ] - ], - trg_sentence: Any, - cepts: Optional[ - Union[ - List[List[Union[int, Any]]], List[List[Any]], List[List[int]] - ] - ], - ) -> None: ... - def center_of_cept(self, i: Optional[int]) -> int: ... - def fertility_of_i(self, i: int) -> int: ... - def is_head_word(self, j: int) -> bool: ... - def previous_cept(self, j: int) -> Optional[int]: ... - def previous_in_tablet(self, j: int) -> int: ... - -class IBMModel: - def __init__( - self, sentence_aligned_corpus: List[Union[AlignedSent, Any]] - ) -> None: ... - def best_model2_alignment( - self, - sentence_pair: AlignedSent, - j_pegged: Optional[int] = ..., - i_pegged: int = ..., - ) -> AlignmentInfo: ... - def hillclimb( - self, alignment_info: AlignmentInfo, j_pegged: Optional[int] = ... - ) -> AlignmentInfo: ... - def init_vocab( - self, sentence_aligned_corpus: List[Union[AlignedSent, Any]] - ) -> Incomplete: ... - def neighboring( - self, alignment_info: AlignmentInfo, j_pegged: Optional[int] = ... - ) -> Set[AlignmentInfo]: ... - def reset_probabilities(self) -> Incomplete: ... - def sample( - self, sentence_pair: AlignedSent - ) -> Tuple[Set[AlignmentInfo], AlignmentInfo]: ... - -class Counts: - t_given_s: Incomplete - any_t_given_s: Incomplete - p0: float - p1: float - fertility: Incomplete - fertility_for_any_phi: Incomplete - def __init__(self) -> None: ... - def update_lexical_translation( - self, count: Incomplete, alignment_info: Incomplete, j: Incomplete - ) -> None: ... - def update_null_generation( - self, count: Incomplete, alignment_info: Incomplete - ) -> None: ... - def update_fertility( - self, count: Incomplete, alignment_info: Incomplete - ) -> None: ... diff --git a/stubs/nltk/translate/meteor_score.pyi b/stubs/nltk/translate/meteor_score.pyi deleted file mode 100644 index 6d79c8c..0000000 --- a/stubs/nltk/translate/meteor_score.pyi +++ /dev/null @@ -1,93 +0,0 @@ -from itertools import product as product -from typing import ( - Callable, - Iterable, - List, - Tuple, -) - -from nltk.corpus import ( - WordNetCorpusReader as WordNetCorpusReader, -) -from nltk.corpus import ( - wordnet as wordnet, -) -from nltk.stem.api import StemmerI -from nltk.stem.porter import PorterStemmer as PorterStemmer - -def _count_chunks(matches: List[Tuple[int, int]]) -> int: ... -def _enum_align_words( - enum_hypothesis_list: List[Tuple[int, str]], - enum_reference_list: List[Tuple[int, str]], - stemmer: StemmerI = ..., - wordnet: WordNetCorpusReader = ..., -) -> Tuple[ - List[Tuple[int, int]], List[Tuple[int, str]], List[Tuple[int, str]] -]: ... -def _enum_stem_match( - enum_hypothesis_list: List[Tuple[int, str]], - enum_reference_list: List[Tuple[int, str]], - stemmer: StemmerI = ..., -) -> Tuple[ - List[Tuple[int, int]], List[Tuple[int, str]], List[Tuple[int, str]] -]: ... -def _enum_wordnetsyn_match( - enum_hypothesis_list: List[Tuple[int, str]], - enum_reference_list: List[Tuple[int, str]], - wordnet: WordNetCorpusReader = ..., -) -> Tuple[ - List[Tuple[int, int]], List[Tuple[int, str]], List[Tuple[int, str]] -]: ... -def _match_enums( - enum_hypothesis_list: List[Tuple[int, str]], - enum_reference_list: List[Tuple[int, str]], -) -> Tuple[ - List[Tuple[int, int]], List[Tuple[int, str]], List[Tuple[int, str]] -]: ... -def exact_match( - hypothesis: Iterable[str], reference: Iterable[str] -) -> tuple[ - list[tuple[int, int]], list[tuple[int, str]], list[tuple[int, str]] -]: ... -def stem_match( - hypothesis: Iterable[str], - reference: Iterable[str], - stemmer: StemmerI = ..., -) -> tuple[ - list[tuple[int, int]], list[tuple[int, str]], list[tuple[int, str]] -]: ... -def wordnetsyn_match( - hypothesis: Iterable[str], - reference: Iterable[str], - wordnet: WordNetCorpusReader = ..., -) -> tuple[ - list[tuple[int, int]], list[tuple[int, str]], list[tuple[int, str]] -]: ... -def align_words( - hypothesis: Iterable[str], - reference: Iterable[str], - stemmer: StemmerI = ..., - wordnet: WordNetCorpusReader = ..., -) -> tuple[ - list[tuple[int, int]], list[tuple[int, str]], list[tuple[int, str]] -]: ... -def single_meteor_score( - reference: Iterable[str], - hypothesis: Iterable[str], - preprocess: Callable[[str], str] = ..., - stemmer: StemmerI = ..., - wordnet: WordNetCorpusReader = ..., - alpha: float = 0.9, - beta: float = 3.0, - gamma: float = 0.5, -) -> float: ... -def meteor_score( - references: Iterable[Iterable[str]], - hypothesis: Iterable[str], - preprocess: Callable[[str], str] = ..., - stemmer: StemmerI = ..., - wordnet: WordNetCorpusReader = ..., - alpha: float = 0.9, - beta: float = 3.0, - gamma: float = 0.5, -) -> float: ... diff --git a/stubs/nltk/translate/metrics.pyi b/stubs/nltk/translate/metrics.pyi deleted file mode 100644 index 9feeca9..0000000 --- a/stubs/nltk/translate/metrics.pyi +++ /dev/null @@ -1,7 +0,0 @@ -from _typeshed import Incomplete - -def alignment_error_rate( - reference: Incomplete, - hypothesis: Incomplete, - possible: Incomplete | None = None, -) -> Incomplete: ... diff --git a/stubs/nltk/translate/nist_score.pyi b/stubs/nltk/translate/nist_score.pyi deleted file mode 100644 index d53b92c..0000000 --- a/stubs/nltk/translate/nist_score.pyi +++ /dev/null @@ -1,15 +0,0 @@ -from typing import List - -from _typeshed import Incomplete - -from nltk.util import ngrams as ngrams - -def corpus_nist( - list_of_references: List[List[List[str]]], - hypotheses: List[List[str]], - n: int = ..., -) -> float: ... -def nist_length_penalty(ref_len: int, hyp_len: int) -> float: ... -def sentence_nist( - references: Incomplete, hypothesis: Incomplete, n: int = 5 -) -> Incomplete: ... diff --git a/stubs/nltk/translate/phrase_based.pyi b/stubs/nltk/translate/phrase_based.pyi deleted file mode 100644 index 97b4db5..0000000 --- a/stubs/nltk/translate/phrase_based.pyi +++ /dev/null @@ -1,21 +0,0 @@ -from _typeshed import Incomplete - -def extract( - f_start: Incomplete, - f_end: Incomplete, - e_start: Incomplete, - e_end: Incomplete, - alignment: Incomplete, - f_aligned: Incomplete, - srctext: Incomplete, - trgtext: Incomplete, - srclen: Incomplete, - trglen: Incomplete, - max_phrase_length: Incomplete, -) -> Incomplete: ... -def phrase_extraction( - srctext: Incomplete, - trgtext: Incomplete, - alignment: Incomplete, - max_phrase_length: int = 0, -) -> Incomplete: ... diff --git a/stubs/nltk/translate/ribes_score.pyi b/stubs/nltk/translate/ribes_score.pyi deleted file mode 100644 index 018eba6..0000000 --- a/stubs/nltk/translate/ribes_score.pyi +++ /dev/null @@ -1,40 +0,0 @@ -from typing import ( - Any, - Iterator, - List, - Tuple, - Union, -) - -from nltk.util import choose as choose -from nltk.util import ngrams as ngrams - -def corpus_ribes( - list_of_references: List[List[List[str]]], - hypotheses: List[List[str]], - alpha: float = ..., - beta: float = ..., -) -> float: ... -def find_increasing_sequences( - worder: List[int], -) -> Iterator[ - Union[ - Tuple[int, int], - Tuple[int, int, int], - Tuple[int, int, int, int, int, int], - Tuple[int, int, int, int], - ] -]: ... -def kendall_tau( - worder: List[Union[int, Any]], normalize: bool = ... -) -> float: ... -def position_of_ngram(ngram: Tuple[str, str], sentence: List[str]) -> int: ... -def sentence_ribes( - references: List[List[str]], - hypothesis: List[str], - alpha: float = ..., - beta: float = ..., -) -> float: ... -def word_rank_alignment( - reference: List[str], hypothesis: List[str], character_based: bool = ... -) -> List[Union[int, Any]]: ... diff --git a/stubs/nltk/translate/stack_decoder.pyi b/stubs/nltk/translate/stack_decoder.pyi deleted file mode 100644 index e518453..0000000 --- a/stubs/nltk/translate/stack_decoder.pyi +++ /dev/null @@ -1,61 +0,0 @@ -from typing import ( - Any, - DefaultDict, - List, - Optional, - Tuple, - Union, -) - -from _typeshed import Incomplete - -from nltk.translate.api import PhraseTable - -class StackDecoder: - def __init__( - self, phrase_table: Optional[PhraseTable], language_model: None - ) -> None: ... - def compute_future_scores( - self, src_sentence: Tuple[str, str, str, str, str, str] - ) -> DefaultDict[int, DefaultDict[int, float]]: ... - def distortion_score( - self, hypothesis: _Hypothesis, next_src_phrase_span: Tuple[int, int] - ) -> float: ... - def find_all_src_phrases( - self, src_sentence: Tuple[str, str, str, str, str, str] - ) -> List[List[Union[int, Any]]]: ... - def future_score( - self, - hypothesis: _Hypothesis, - future_score_table: DefaultDict[int, DefaultDict[int, float]], - sentence_length: int, - ) -> float: ... - @staticmethod - def valid_phrases( - all_phrases_from: List[List[Union[int, Any]]], hypothesis: _Hypothesis - ) -> List[Tuple[int, int]]: ... - -class _Hypothesis: - def __init__( - self, - raw_score: float = ..., - src_phrase_span: Tuple[()] = ..., - trg_phrase: Tuple[()] = ..., - previous: None = ..., - future_score: float = ..., - ) -> None: ... - def score(self) -> float: ... - def translated_positions(self) -> List[Any]: ... - def translation_so_far(self) -> List[Any]: ... - def untranslated_spans( - self, sentence_length: int - ) -> List[Tuple[int, int]]: ... - -class _Stack: - def __contains__(self, hypothesis: _Hypothesis) -> bool: ... - def __init__( - self, max_size: int = ..., beam_threshold: float = ... - ) -> None: ... - def best(self) -> _Hypothesis: ... - def push(self, hypothesis: _Hypothesis) -> Incomplete: ... - def threshold_prune(self) -> Incomplete: ... diff --git a/stubs/nltk/tree/__init__.pyi b/stubs/nltk/tree/__init__.pyi deleted file mode 100644 index a0ca666..0000000 --- a/stubs/nltk/tree/__init__.pyi +++ /dev/null @@ -1,53 +0,0 @@ -from nltk.tree.immutable import ( - ImmutableMultiParentedTree as ImmutableMultiParentedTree, -) -from nltk.tree.immutable import ( - ImmutableParentedTree as ImmutableParentedTree, -) -from nltk.tree.immutable import ( - ImmutableProbabilisticTree as ImmutableProbabilisticTree, -) -from nltk.tree.immutable import ( - ImmutableTree as ImmutableTree, -) -from nltk.tree.parented import ( - MultiParentedTree as MultiParentedTree, -) -from nltk.tree.parented import ( - ParentedTree as ParentedTree, -) -from nltk.tree.parsing import ( - bracket_parse as bracket_parse, -) -from nltk.tree.parsing import ( - sinica_parse as sinica_parse, -) -from nltk.tree.prettyprinter import TreePrettyPrinter as TreePrettyPrinter -from nltk.tree.probabilistic import ProbabilisticTree as ProbabilisticTree -from nltk.tree.transforms import ( - chomsky_normal_form as chomsky_normal_form, -) -from nltk.tree.transforms import ( - collapse_unary as collapse_unary, -) -from nltk.tree.transforms import ( - un_chomsky_normal_form as un_chomsky_normal_form, -) -from nltk.tree.tree import Tree as Tree - -__all__ = [ - "ImmutableMultiParentedTree", - "ImmutableParentedTree", - "ImmutableProbabilisticTree", - "ImmutableTree", - "MultiParentedTree", - "ParentedTree", - "bracket_parse", - "sinica_parse", - "TreePrettyPrinter", - "ProbabilisticTree", - "chomsky_normal_form", - "collapse_unary", - "un_chomsky_normal_form", - "Tree", -] diff --git a/stubs/nltk/tree/immutable.pyi b/stubs/nltk/tree/immutable.pyi deleted file mode 100644 index 9211300..0000000 --- a/stubs/nltk/tree/immutable.pyi +++ /dev/null @@ -1,47 +0,0 @@ -from _typeshed import Incomplete - -from nltk.probability import ProbabilisticMixIn -from nltk.tree.parented import MultiParentedTree, ParentedTree -from nltk.tree.tree import Tree - -__all__ = [ - "ImmutableProbabilisticTree", - "ImmutableTree", - "ImmutableParentedTree", - "ImmutableMultiParentedTree", -] - -class ImmutableTree(Tree): - def __init__( - self, node: Incomplete, children: Incomplete | None = None - ) -> None: ... - def __setitem__(self, index: Incomplete, value: Incomplete) -> None: ... - def __setslice__( - self, i: Incomplete, j: Incomplete, value: Incomplete - ) -> None: ... - def __delitem__(self, index: Incomplete) -> None: ... - def __delslice__(self, i: Incomplete, j: Incomplete) -> None: ... - def __iadd__(self, other: Incomplete) -> None: ... - def __imul__(self, other: Incomplete) -> None: ... - def append(self, v: Incomplete) -> None: ... - def extend(self, v: Incomplete) -> None: ... - def pop(self, v: Incomplete | None = None) -> None: ... - def remove(self, v: Incomplete) -> None: ... - def reverse(self) -> None: ... - def sort(self) -> None: ... - def __hash__(self) -> Incomplete: ... - def set_label(self, value: Incomplete) -> None: ... - -class ImmutableProbabilisticTree(ImmutableTree, ProbabilisticMixIn): - def __init__( - self, - node: Incomplete, - children: Incomplete | None = None, - **prob_kwargs: Incomplete, - ) -> None: ... - def copy(self, deep: bool = False) -> Incomplete: ... - @classmethod - def convert(cls: Incomplete, val: Incomplete) -> Incomplete: ... - -class ImmutableParentedTree(ImmutableTree, ParentedTree): ... -class ImmutableMultiParentedTree(ImmutableTree, MultiParentedTree): ... diff --git a/stubs/nltk/tree/parented.pyi b/stubs/nltk/tree/parented.pyi deleted file mode 100644 index 03828dd..0000000 --- a/stubs/nltk/tree/parented.pyi +++ /dev/null @@ -1,51 +0,0 @@ -from typing import ( - List, - Optional, - Tuple, - Union, -) - -from _typeshed import Incomplete - -__all__ = ["ParentedTree", "MultiParentedTree"] - -class AbstractParentedTree: - def __init__( - self, - node: str, - children: Optional[List[Union[str, ParentedTree]]] = ..., - ) -> None: ... - -class ParentedTree: - def __init__( - self, - node: str, - children: Optional[List[Union[str, ParentedTree]]] = ..., - ) -> None: ... - def _setparent( - self, child: ParentedTree, index: int, dry_run: bool = ... - ) -> Incomplete: ... - def left_sibling(self) -> ParentedTree: ... - def parent(self) -> Optional[ParentedTree]: ... - def parent_index(self) -> int: ... - def right_sibling(self) -> ParentedTree: ... - def root(self) -> ParentedTree: ... - def treeposition( - self, - ) -> Union[ - Tuple[int, int, int, int], - Tuple[int, int], - Tuple[int, int, int], - Tuple[int], - ]: ... - -class MultiParentedTree(AbstractParentedTree): - def __init__( - self, node: Incomplete, children: Incomplete | None = None - ) -> None: ... - def parents(self) -> Incomplete: ... - def left_siblings(self) -> Incomplete: ... - def right_siblings(self) -> Incomplete: ... - def roots(self) -> Incomplete: ... - def parent_indices(self, parent: Incomplete) -> Incomplete: ... - def treepositions(self, root: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tree/parsing.pyi b/stubs/nltk/tree/parsing.pyi deleted file mode 100644 index 7b92fb8..0000000 --- a/stubs/nltk/tree/parsing.pyi +++ /dev/null @@ -1,5 +0,0 @@ -__all__ = ["bracket_parse", "sinica_parse"] -from _typeshed import Incomplete - -def bracket_parse(s: Incomplete) -> None: ... -def sinica_parse(s: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tree/prettyprinter.pyi b/stubs/nltk/tree/prettyprinter.pyi deleted file mode 100644 index f5ecda7..0000000 --- a/stubs/nltk/tree/prettyprinter.pyi +++ /dev/null @@ -1,33 +0,0 @@ -from _typeshed import Incomplete - -__all__ = ["TreePrettyPrinter"] - -class TreePrettyPrinter: - def __init__( - self, - tree: Incomplete, - sentence: Incomplete | None = None, - highlight: Incomplete = (), - ) -> None: ... - @staticmethod - def nodecoords( - tree: Incomplete, sentence: Incomplete, highlight: Incomplete - ) -> Incomplete: ... - def text( - self, - nodedist: int = 1, - unicodelines: bool = False, - html: bool = False, - ansi: bool = False, - nodecolor: str = "blue", - leafcolor: str = "red", - funccolor: str = "green", - abbreviate: Incomplete | None = None, - maxwidth: int = 16, - ) -> Incomplete: ... - def svg( - self, - nodecolor: str = "blue", - leafcolor: str = "red", - funccolor: str = "green", - ) -> Incomplete: ... diff --git a/stubs/nltk/tree/probabilistic.pyi b/stubs/nltk/tree/probabilistic.pyi deleted file mode 100644 index b40a260..0000000 --- a/stubs/nltk/tree/probabilistic.pyi +++ /dev/null @@ -1,19 +0,0 @@ -from _typeshed import Incomplete - -from nltk.probability import ProbabilisticMixIn -from nltk.tree.tree import Tree - -__all__ = ["ProbabilisticTree"] - -class ProbabilisticTree(Tree, ProbabilisticMixIn): - def __init__( - self, - node: Incomplete, - children: Incomplete | None = None, - **prob_kwargs: Incomplete, - ) -> None: ... - def copy(self, deep: bool = False) -> Incomplete: ... - @classmethod - def convert(cls: Incomplete, val: Incomplete) -> Incomplete: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/tree/transforms.pyi b/stubs/nltk/tree/transforms.pyi deleted file mode 100644 index 44f9930..0000000 --- a/stubs/nltk/tree/transforms.pyi +++ /dev/null @@ -1,25 +0,0 @@ -from _typeshed import Incomplete - -__all__ = ["chomsky_normal_form", "un_chomsky_normal_form", "collapse_unary"] - -def chomsky_normal_form( - tree: Incomplete, - factor: str = "right", - horzMarkov: Incomplete | None = None, - vertMarkov: int = 0, - childChar: str = "|", - parentChar: str = "^", -) -> None: ... -def un_chomsky_normal_form( - tree: Incomplete, - expandUnary: bool = True, - childChar: str = "|", - parentChar: str = "^", - unaryChar: str = "+", -) -> None: ... -def collapse_unary( - tree: Incomplete, - collapsePOS: bool = False, - collapseRoot: bool = False, - joinChar: str = "+", -) -> None: ... diff --git a/stubs/nltk/tree/tree.pyi b/stubs/nltk/tree/tree.pyi deleted file mode 100644 index 85b70c9..0000000 --- a/stubs/nltk/tree/tree.pyi +++ /dev/null @@ -1,98 +0,0 @@ -from collections.abc import Generator -from typing import Any - -from _typeshed import Incomplete - -__all__ = ["Tree"] - -class Tree(list): # type: ignore[type-arg] - def __init__( - self, node: Incomplete, children: Incomplete | None = None - ) -> None: ... - def __eq__(self, other: Incomplete) -> Incomplete: ... - def __lt__(self, other: Incomplete) -> Incomplete: ... - __ne__: Incomplete - __gt__: Incomplete - __le__: Incomplete - __ge__: Incomplete - def __mul__(self, v: Incomplete) -> list[Any]: ... - def __rmul__(self, v: Incomplete) -> list[Any]: ... - def __add__(self, v: Incomplete) -> None: ... - def __radd__(self, v: Incomplete) -> None: ... - def __getitem__(self, index: Incomplete) -> Incomplete: ... - def __setitem__(self, index: Incomplete, value: Incomplete) -> None: ... - def __delitem__(self, index: Incomplete) -> None: ... - node: Incomplete - def label(self) -> Incomplete: ... - def set_label(self, label: Incomplete) -> None: ... - def leaves(self) -> Incomplete: ... - def flatten(self) -> Incomplete: ... - def height(self) -> Incomplete: ... - def treepositions(self, order: str = "preorder") -> Incomplete: ... - def subtrees( - self, filter: Incomplete | None = None - ) -> Generator[Incomplete, Incomplete, None]: ... - def productions(self) -> Incomplete: ... - def pos(self) -> Incomplete: ... - def leaf_treeposition(self, index: Incomplete) -> Incomplete: ... - def treeposition_spanning_leaves( - self, start: Incomplete, end: Incomplete - ) -> Incomplete: ... - def chomsky_normal_form( - self, - factor: str = "right", - horzMarkov: Incomplete | None = None, - vertMarkov: int = 0, - childChar: str = "|", - parentChar: str = "^", - ) -> None: ... - def un_chomsky_normal_form( - self, - expandUnary: bool = True, - childChar: str = "|", - parentChar: str = "^", - unaryChar: str = "+", - ) -> None: ... - def collapse_unary( - self, - collapsePOS: bool = False, - collapseRoot: bool = False, - joinChar: str = "+", - ) -> None: ... - @classmethod - def convert(cls: Incomplete, tree: Incomplete) -> Incomplete: ... - def __copy__(self) -> Incomplete: ... - def __deepcopy__(self, memo: Incomplete) -> Incomplete: ... - def copy(self, deep: bool = False) -> Incomplete: ... - def freeze(self, leaf_freezer: Incomplete | None = None) -> Incomplete: ... - @classmethod - def fromstring( - cls: Incomplete, - s: Incomplete, - brackets: str = "()", - read_node: Incomplete | None = None, - read_leaf: Incomplete | None = None, - node_pattern: Incomplete | None = None, - leaf_pattern: Incomplete | None = None, - remove_empty_top_bracketing: bool = False, - ) -> Incomplete: ... - @classmethod - def fromlist(cls: Incomplete, l: Incomplete) -> Incomplete: ... - def draw(self) -> None: ... - def pretty_print( - self, - sentence: Incomplete | None = None, - highlight: Incomplete = (), - stream: Incomplete | None = None, - **kwargs: Incomplete, - ) -> None: ... - def pprint(self, **kwargs: Incomplete) -> None: ... - def pformat( - self, - margin: int = 70, - indent: int = 0, - nodesep: str = "", - parens: str = "()", - quotes: bool = False, - ) -> Incomplete: ... - def pformat_latex_qtree(self) -> Incomplete: ... diff --git a/stubs/nltk/treeprettyprinter.pyi b/stubs/nltk/treeprettyprinter.pyi deleted file mode 100644 index 4424b37..0000000 --- a/stubs/nltk/treeprettyprinter.pyi +++ /dev/null @@ -1,6 +0,0 @@ -from nltk.internals import Deprecated -from nltk.tree.prettyprinter import TreePrettyPrinter as TPP - -__all__ = ["TreePrettyPrinter"] - -class TreePrettyPrinter(Deprecated, TPP): ... diff --git a/stubs/nltk/treetransforms.pyi b/stubs/nltk/treetransforms.pyi deleted file mode 100644 index f659ced..0000000 --- a/stubs/nltk/treetransforms.pyi +++ /dev/null @@ -1,7 +0,0 @@ -from _typeshed import Incomplete - -__all__ = ["chomsky_normal_form", "un_chomsky_normal_form", "collapse_unary"] - -chomsky_normal_form: Incomplete -un_chomsky_normal_form: Incomplete -collapse_unary: Incomplete diff --git a/stubs/nltk/twitter/__init__.pyi b/stubs/nltk/twitter/__init__.pyi deleted file mode 100644 index caba79b..0000000 --- a/stubs/nltk/twitter/__init__.pyi +++ /dev/null @@ -1,22 +0,0 @@ -from nltk.twitter.common import json2csv as json2csv -from nltk.twitter.twitterclient import ( - Query as Query, -) -from nltk.twitter.twitterclient import ( - Streamer as Streamer, -) -from nltk.twitter.twitterclient import ( - TweetViewer as TweetViewer, -) -from nltk.twitter.twitterclient import ( - TweetWriter as TweetWriter, -) -from nltk.twitter.twitterclient import ( - Twitter as Twitter, -) -from nltk.twitter.util import ( - Authenticate as Authenticate, -) -from nltk.twitter.util import ( - credsfromfile as credsfromfile, -) diff --git a/stubs/nltk/twitter/api.pyi b/stubs/nltk/twitter/api.pyi deleted file mode 100644 index 95b57d3..0000000 --- a/stubs/nltk/twitter/api.pyi +++ /dev/null @@ -1,39 +0,0 @@ -from abc import ABCMeta, abstractmethod -from datetime import tzinfo - -from _typeshed import Incomplete - -class LocalTimezoneOffsetWithUTC(tzinfo): # type: ignore[misc] - STDOFFSET: Incomplete - DSTOFFSET: Incomplete - DSTOFFSET = STDOFFSET - def utcoffset(self, dt: Incomplete) -> Incomplete: ... - -LOCAL: Incomplete - -class BasicTweetHandler(metaclass=ABCMeta): - limit: Incomplete - counter: int - do_stop: bool - max_id: Incomplete - def __init__(self, limit: int = 20) -> None: ... - def do_continue(self) -> Incomplete: ... - -class TweetHandlerI(BasicTweetHandler): - upper_date_limit: Incomplete - lower_date_limit: Incomplete - startingup: bool - def __init__( - self, - limit: int = 20, - upper_date_limit: Incomplete | None = None, - lower_date_limit: Incomplete | None = None, - ) -> None: ... - @abstractmethod - def handle(self, data: Incomplete) -> Incomplete: ... - @abstractmethod - def on_finish(self) -> Incomplete: ... - do_stop: bool - def check_date_limit( - self, data: Incomplete, verbose: bool = False - ) -> None: ... diff --git a/stubs/nltk/twitter/common.pyi b/stubs/nltk/twitter/common.pyi deleted file mode 100644 index 3c710de..0000000 --- a/stubs/nltk/twitter/common.pyi +++ /dev/null @@ -1,34 +0,0 @@ -from _typeshed import Incomplete - -from nltk.internals import deprecated as deprecated - -HIER_SEPARATOR: str - -def extract_fields(tweet: Incomplete, fields: Incomplete) -> Incomplete: ... -def json2csv( - fp: Incomplete, - outfile: Incomplete, - fields: Incomplete, - encoding: str = "utf8", - errors: str = "replace", - gzip_compress: bool = False, -) -> None: ... -def outf_writer_compat( - outfile: Incomplete, - encoding: Incomplete, - errors: Incomplete, - gzip_compress: bool = False, -) -> Incomplete: ... -def json2csv_entities( - tweets_file: Incomplete, - outfile: Incomplete, - main_fields: Incomplete, - entity_type: Incomplete, - entity_fields: Incomplete, - encoding: str = "utf8", - errors: str = "replace", - gzip_compress: bool = False, -) -> None: ... -def get_header_field_list( - main_fields: Incomplete, entity_type: Incomplete, entity_fields: Incomplete -) -> Incomplete: ... diff --git a/stubs/nltk/twitter/twitter_demo.pyi b/stubs/nltk/twitter/twitter_demo.pyi deleted file mode 100644 index 11e2fd3..0000000 --- a/stubs/nltk/twitter/twitter_demo.pyi +++ /dev/null @@ -1,42 +0,0 @@ -from _typeshed import Incomplete - -from nltk.twitter import ( - Query as Query, -) -from nltk.twitter import ( - Streamer as Streamer, -) -from nltk.twitter import ( - TweetViewer as TweetViewer, -) -from nltk.twitter import ( - TweetWriter as TweetWriter, -) -from nltk.twitter import ( - Twitter as Twitter, -) -from nltk.twitter import ( - credsfromfile as credsfromfile, -) - -SPACER: str - -def verbose(func: Incomplete) -> Incomplete: ... -def yesterday() -> Incomplete: ... -def setup() -> None: ... -def twitterclass_demo() -> None: ... -def sampletoscreen_demo(limit: int = 20) -> None: ... -def tracktoscreen_demo( - track: str = "taylor swift", limit: int = 10 -) -> None: ... -def search_demo(keywords: str = "nltk") -> None: ... -def tweets_by_user_demo(user: str = "NLTK_org", count: int = 200) -> None: ... -def lookup_by_userid_demo() -> None: ... -def followtoscreen_demo(limit: int = 10) -> None: ... -def streamtofile_demo(limit: int = 20) -> None: ... -def limit_by_time_demo(keywords: str = "nltk") -> None: ... -def corpusreader_demo() -> None: ... -def expand_tweetids_demo() -> None: ... - -ALL: Incomplete -DEMOS: Incomplete diff --git a/stubs/nltk/twitter/twitterclient.pyi b/stubs/nltk/twitter/twitterclient.pyi deleted file mode 100644 index 0328313..0000000 --- a/stubs/nltk/twitter/twitterclient.pyi +++ /dev/null @@ -1,109 +0,0 @@ -from collections.abc import Generator - -from _typeshed import Incomplete -from twython import Twython, TwythonStreamer # type: ignore[import-not-found] - -from nltk.twitter.api import ( - BasicTweetHandler as BasicTweetHandler, -) -from nltk.twitter.api import ( - TweetHandlerI as TweetHandlerI, -) -from nltk.twitter.util import ( - credsfromfile as credsfromfile, -) -from nltk.twitter.util import ( - guess_path as guess_path, -) - -class Streamer(TwythonStreamer): # type: ignore[misc] - handler: Incomplete - do_continue: bool - def __init__( - self, - app_key: Incomplete, - app_secret: Incomplete, - oauth_token: Incomplete, - oauth_token_secret: Incomplete, - ) -> None: ... - def register(self, handler: Incomplete) -> None: ... - def on_success(self, data: Incomplete) -> None: ... - def on_error(self, status_code: Incomplete, data: Incomplete) -> None: ... - def sample(self) -> None: ... - def filter( - self, track: str = "", follow: str = "", lang: str = "en" - ) -> None: ... - -class Query(Twython): # type: ignore[misc] - handler: Incomplete - do_continue: bool - def __init__( - self, - app_key: Incomplete, - app_secret: Incomplete, - oauth_token: Incomplete, - oauth_token_secret: Incomplete, - ) -> None: ... - def register(self, handler: Incomplete) -> None: ... - def expand_tweetids( - self, ids_f: Incomplete, verbose: bool = True - ) -> Incomplete: ... - def search_tweets( - self, - keywords: Incomplete, - limit: int = 100, - lang: str = "en", - max_id: Incomplete | None = None, - retries_after_twython_exception: int = 0, - ) -> Generator[Incomplete, None, None]: ... - def user_info_from_id(self, userids: Incomplete) -> Incomplete: ... - def user_tweets( - self, - screen_name: Incomplete, - limit: Incomplete, - include_rts: str = "false", - ) -> None: ... - -class Twitter: - streamer: Incomplete - query: Incomplete - def __init__(self) -> None: ... - def tweets( - self, - keywords: str = "", - follow: str = "", - to_screen: bool = True, - stream: bool = True, - limit: int = 100, - date_limit: Incomplete | None = None, - lang: str = "en", - repeat: bool = False, - gzip_compress: bool = False, - ) -> None: ... - -class TweetViewer(TweetHandlerI): - def handle(self, data: Incomplete) -> None: ... - def on_finish(self) -> None: ... - -class TweetWriter(TweetHandlerI): - fprefix: Incomplete - subdir: Incomplete - gzip_compress: Incomplete - fname: Incomplete - repeat: Incomplete - output: Incomplete - def __init__( - self, - limit: int = 2000, - upper_date_limit: Incomplete | None = None, - lower_date_limit: Incomplete | None = None, - fprefix: str = "tweets", - subdir: str = "twitter-files", - repeat: bool = False, - gzip_compress: bool = False, - ) -> None: ... - def timestamped_file(self) -> Incomplete: ... - startingup: bool - def handle(self, data: Incomplete) -> None: ... - def on_finish(self) -> None: ... - def do_continue(self) -> Incomplete: ... diff --git a/stubs/nltk/twitter/util.pyi b/stubs/nltk/twitter/util.pyi deleted file mode 100644 index effc59e..0000000 --- a/stubs/nltk/twitter/util.pyi +++ /dev/null @@ -1,24 +0,0 @@ -from _typeshed import Incomplete - -def credsfromfile( - creds_file: Incomplete | None = None, - subdir: Incomplete | None = None, - verbose: bool = False, -) -> Incomplete: ... - -class Authenticate: - creds_file: str - creds_fullpath: Incomplete - oauth: Incomplete - twitter_dir: Incomplete - creds_subdir: Incomplete - def __init__(self) -> None: ... - def load_creds( - self, - creds_file: Incomplete | None = None, - subdir: Incomplete | None = None, - verbose: bool = False, - ) -> Incomplete: ... - -def add_access_token(creds_file: Incomplete | None = None) -> None: ... -def guess_path(pth: Incomplete) -> Incomplete: ... diff --git a/stubs/nltk/util.pyi b/stubs/nltk/util.pyi deleted file mode 100644 index 859ee0b..0000000 --- a/stubs/nltk/util.pyi +++ /dev/null @@ -1,148 +0,0 @@ -from collections import defaultdict -from collections.abc import Generator -from itertools import chain -from typing import ( - Any, - Iterator, - List, - Optional, - Tuple, - Union, -) - -from _typeshed import Incomplete - -from nltk.collections import * # type: ignore[import-untyped] -from nltk.internals import ( - deprecated as deprecated, -) -from nltk.internals import ( - raise_unorderable_types as raise_unorderable_types, -) -from nltk.internals import ( - slice_bounds as slice_bounds, -) - -def choose(n: int, k: int) -> int: ... -def everygrams( - sequence: List[str], - min_len: int = ..., - max_len: int = ..., - pad_left: bool = ..., - pad_right: bool = ..., - **kwargs: Incomplete, -) -> Iterator[Union[Tuple[str], Tuple[str, str]]]: ... -def ngrams( - sequence: List[Union[Tuple[str, int], str]], n: int, **kwargs: Incomplete -) -> Iterator[Any]: ... -def pad_sequence( - sequence: List[str], - n: int, - pad_left: bool = ..., - pad_right: bool = ..., - left_pad_symbol: Optional[str] = ..., - right_pad_symbol: Optional[str] = ..., -) -> chain: ... # type: ignore[type-arg] -def unique_list(xs: List[str]) -> List[str]: ... -def usage(obj: Incomplete) -> None: ... -def in_idle() -> Incomplete: ... -def pr( - data: Incomplete, start: int = 0, end: Incomplete | None = None -) -> None: ... -def print_string(s: Incomplete, width: int = 70) -> None: ... -def tokenwrap( - tokens: Incomplete, separator: str = " ", width: int = 70 -) -> Incomplete: ... -def cut_string(s: Incomplete, width: int = 70) -> Incomplete: ... - -class Index(defaultdict): # type: ignore[type-arg] - def __init__(self, pairs: Incomplete) -> None: ... - -def re_show( - regexp: Incomplete, string: Incomplete, left: str = "{", right: str = "}" -) -> None: ... -def filestring(f: Incomplete) -> Incomplete: ... -def breadth_first( - tree: Incomplete, children: Incomplete = ..., maxdepth: int = -1 -) -> Generator[Incomplete, None, None]: ... -def edge_closure( - tree: Incomplete, - children: Incomplete = ..., - maxdepth: int = -1, - verbose: bool = False, -) -> Generator[Incomplete, None, None]: ... -def edges2dot( - edges: Incomplete, - shapes: Incomplete | None = None, - attr: Incomplete | None = None, -) -> Incomplete: ... -def unweighted_minimum_spanning_digraph( - tree: Incomplete, - children: Incomplete = ..., - shapes: Incomplete | None = None, - attr: Incomplete | None = None, -) -> Incomplete: ... -def acyclic_breadth_first( - tree: Incomplete, - children: Incomplete = ..., - maxdepth: int = -1, - verbose: bool = False, -) -> Generator[Incomplete, None, None]: ... -def acyclic_depth_first( - tree: Incomplete, - children: Incomplete = ..., - depth: int = -1, - cut_mark: Incomplete | None = None, - traversed: Incomplete | None = None, - verbose: bool = False, -) -> Incomplete: ... -def acyclic_branches_depth_first( - tree: Incomplete, - children: Incomplete = ..., - depth: int = -1, - cut_mark: Incomplete | None = None, - traversed: Incomplete | None = None, - verbose: bool = False, -) -> Incomplete: ... -def acyclic_dic2tree(node: Incomplete, dic: Incomplete) -> Incomplete: ... -def unweighted_minimum_spanning_dict( - tree: Incomplete, children: Incomplete = ... -) -> Incomplete: ... -def unweighted_minimum_spanning_tree( - tree: Incomplete, children: Incomplete = ... -) -> Incomplete: ... -def guess_encoding(data: Incomplete) -> Incomplete: ... -def invert_dict(d: Incomplete) -> Incomplete: ... -def transitive_closure( - graph: Incomplete, reflexive: bool = False -) -> Incomplete: ... -def invert_graph(graph: Incomplete) -> Incomplete: ... -def clean_html(html: Incomplete) -> None: ... -def clean_url(url: Incomplete) -> None: ... -def flatten(*args: Incomplete) -> Incomplete: ... -def bigrams( - sequence: Incomplete, **kwargs: Incomplete -) -> Generator[Incomplete, Incomplete, None]: ... -def trigrams( - sequence: Incomplete, **kwargs: Incomplete -) -> Generator[Incomplete, Incomplete, None]: ... -def skipgrams( - sequence: Incomplete, n: Incomplete, k: Incomplete, **kwargs: Incomplete -) -> Generator[Incomplete, None, None]: ... -def binary_search_file( - file: Incomplete, - key: Incomplete, - cache: Incomplete | None = None, - cacheDepth: int = -1, -) -> Incomplete: ... -def set_proxy( - proxy: Incomplete, user: Incomplete | None = None, password: str = "" -) -> None: ... -def elementtree_indent(elem: Incomplete, level: int = 0) -> None: ... -def pairwise(iterable: Incomplete) -> Incomplete: ... -def parallelize_preprocess( - func: Incomplete, - iterator: Incomplete, - processes: Incomplete, - progress_bar: bool = False, -) -> Incomplete: ... diff --git a/stubs/nltk/wsd.pyi b/stubs/nltk/wsd.pyi deleted file mode 100644 index 49b313c..0000000 --- a/stubs/nltk/wsd.pyi +++ /dev/null @@ -1,11 +0,0 @@ -from _typeshed import Incomplete - -from nltk.corpus import wordnet as wordnet - -def lesk( - context_sentence: Incomplete, - ambiguous_word: Incomplete, - pos: Incomplete | None = None, - synsets: Incomplete | None = None, - lang: str = "eng", -) -> Incomplete: ... diff --git a/tox.ini b/tox.ini index bb49f2f..a27ecb8 100644 --- a/tox.ini +++ b/tox.ini @@ -39,7 +39,7 @@ commands = stubtest dill --mypy-config-file mypy.ini --allowlist stubs/dill/allowlist --ignore-unused-allowlist stubtest lemminflect --mypy-config-file mypy.ini --allowlist stubs/lemminflect/allowlist --ignore-unused-allowlist stubtest nltk --mypy-config-file mypy.ini --allowlist stubs/nltk/allowlist --ignore-unused-allowlist - stubtest colors --mypy-config-file mypy.ini + stubtest colors --mypy-config-file mypy.ini --allowlist stubs/colors/allowlist --ignore-unused-allowlist stubtest aenum --mypy-config-file mypy.ini --allowlist stubs/aenum/allowlist --ignore-unused-allowlist [gh]