Skip to content

Commit b80863d

Browse files
committed
new maintainer
1 parent d92b74c commit b80863d

File tree

8 files changed

+415
-253
lines changed

8 files changed

+415
-253
lines changed

‎.Rbuildignore‎

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,3 +13,5 @@
1313
^Makefile$
1414
^_pkgdown\.yml$
1515
^CRAN-SUBMISSION$
16+
^Dockerfile$
17+
^docker-compose.yml$

‎DESCRIPTION‎

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
11
Package: tokenizers
22
Type: Package
33
Title: Fast, Consistent Tokenization of Natural Language Text
4-
Version: 0.3.0
5-
Date: 2022-12-19
4+
Version: 0.3.1
5+
Date: 2024-03-27
66
Description: Convert natural language text into tokens. Includes tokenizers for
77
shingled n-grams, skip n-grams, words, word stems, sentences, paragraphs,
88
characters, shingled characters, lines, Penn Treebank, regular
@@ -13,7 +13,10 @@ Description: Convert natural language text into tokens. Includes tokenizers for
1313
yet correct tokenization in 'UTF-8'.
1414
License: MIT + file LICENSE
1515
LazyData: yes
16-
Authors@R: c(person("Lincoln", "Mullen", role = c("aut", "cre"),
16+
Authors@R: c(person("Thomas", "Charlon", role = c("aut", "cre"),
17+
email = "[email protected]",
18+
comment = c(ORCID = "0000-0001-7497-0470")),
19+
person("Lincoln", "Mullen", role = c("aut"),
1720
email = "[email protected]",
1821
comment = c(ORCID = "0000-0001-5103-6917")),
1922
person("Os", "Keyes", role = c("ctb"),
@@ -29,7 +32,7 @@ Authors@R: c(person("Lincoln", "Mullen", role = c("aut", "cre"),
2932
comment = c(ORCID = "0000-0002-0797-564X")))
3033
URL: https://docs.ropensci.org/tokenizers/, https://github.com/ropensci/tokenizers
3134
BugReports: https://github.com/ropensci/tokenizers/issues
32-
RoxygenNote: 7.2.1
35+
RoxygenNote: 7.3.1
3336
Depends:
3437
R (>= 3.1.3)
3538
Imports:

‎Dockerfile‎

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,5 @@
1+
2+
from rocker/shiny-verse:4.3.2
3+
4+
add ./ /tokenizers
5+
run R -e "devtools::install('tokenizers', dependencies = TRUE)"

‎Makefile‎

Lines changed: 44 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,44 @@
1+
# prepare the package for release
2+
PKGNAME := $(shell sed -n "s/Package: *\([^ ]*\)/\1/p" DESCRIPTION)
3+
PKGVERS := $(shell sed -n "s/Version: *\([^ ]*\)/\1/p" DESCRIPTION)
4+
PKGSRC := $(shell basename `pwd`)
5+
6+
all: clean devtools_check
7+
8+
doc.pdf:
9+
R CMD Rd2pdf -o doc.pdf .
10+
11+
build:
12+
cd ..;\
13+
R CMD build --no-manual $(PKGSRC)
14+
15+
build-cran:
16+
cd ..;\
17+
R CMD build $(PKGSRC)
18+
19+
install: build
20+
cd ..;\
21+
R CMD INSTALL $(PKGNAME)_$(PKGVERS).tar.gz
22+
23+
check: build-cran
24+
cd ..;\
25+
R CMD check $(PKGNAME)_$(PKGVERS).tar.gz --as-cran
26+
27+
roxygenise:
28+
R -e "roxygen2::roxygenise()"
29+
30+
devtools_test:
31+
R -e "devtools::test()"
32+
33+
devtools_check:
34+
R -e "devtools::check()"
35+
36+
vignette:
37+
cd vignettes;\
38+
R -e "rmarkdown::render('introduction-to-tokenizers.Rmd')";\
39+
R -e "rmarkdown::render('tif-and-tokenizers.Rmd')"
40+
41+
clean:
42+
$(RM) doc.pdf
43+
cd vignettes;\
44+
$(RM) *.pdf *.aux *.bbl *.blg *.out *.tex *.log

‎docker-compose.yml‎

Lines changed: 33 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,33 @@
1+
version: '2'
2+
3+
services:
4+
manual:
5+
build: ./
6+
volumes:
7+
- ./:/tokenizers/
8+
working_dir: /tokenizers
9+
command: make doc.pdf
10+
vignette:
11+
build: ./
12+
volumes:
13+
- ./:/tokenizers/
14+
working_dir: /tokenizers
15+
command: make vignette
16+
build_doc:
17+
build: ./
18+
volumes:
19+
- ./:/tokenizers/
20+
working_dir: /tokenizers/
21+
command: make roxygenise
22+
pkg_test:
23+
build: ./
24+
volumes:
25+
- ./:/tokenizers/
26+
working_dir: /tokenizers/
27+
command: make devtools_test
28+
pkg_check:
29+
build: ./
30+
volumes:
31+
- ./:/tokenizers/
32+
working_dir: /tokenizers/
33+
command: make

‎man/tokenizers.Rd‎

Lines changed: 27 additions & 0 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

0 commit comments

Comments
 (0)