{"@context":"https://w3id.org/codemeta/3.0","@type":"SoftwareSourceCode","identifier":"pkg:npm/%40novel-segment/stringify","name":"@novel-segment/stringify","description":"Chinese word segmentation 簡繁中文分词模块 以網路小說為樣本 基于 Node.js 的中文分词模块","version":"1.0.17","softwareVersion":"1.0.17","license":"https://spdx.org/licenses/ISC","codeRepository":"https://github.com/bluelovers/ws-segment","issueTracker":"https://github.com/bluelovers/ws-segment/issues","url":"https://github.com/bluelovers/ws-segment/tree/master/packages/@novel-segment/stringify#readme","keywords":["NLP","PanGuSegment","PoS tagging","analyzer","async","chinese","chinese segmentation","data","dict","dictionary","file","hanzi","jieba","load","loader","machine learning","morphological analysis","natural language processing","node-novel","novel","novel-segment","node-segment","opencc","optimizer","pangu","phrases","segment","segmentalization","segmentation","segmenter","split","stopword","stream","sync","synonym","text mining","tokenizer","variant","word","word segmentation","中文","中文分詞","中文分词","分詞","分词","盘古分词","盤古分詞","簡繁","create-by-yarn-tool","create-by-tsdx"],"programmingLanguage":{"@type":"ComputerLanguage","name":"TypeScript"},"maintainer":[{"@type":"Person","name":"bluelovers"}],"author":[{"@type":"Person","name":"bluelovers"}],"copyrightHolder":[{"@type":"Person","name":"bluelovers"}],"dateCreated":"2022-01-24","dateModified":"2024-01-03","datePublished":"2024-01-03","copyrightYear":2022,"downloadUrl":"https://registry.npmjs.org/@novel-segment/stringify/-/stringify-1.0.17.tgz","applicationCategory":"npm","runtimePlatform":"npm","developmentStatus":"active","sameAs":["https://www.npmjs.com/package/@novel-segment/stringify"],"funder":[{"@type":"Organization","url":"https://github.com/sponsors/bluelovers"}],"https://www.w3.org/ns/activitystreams#likes":11}