About
Docs
Files
Support
Index

Essentials
AutoDocs
Library functions
Cooperate
Process mastering
Debug
Own debugging kit
FGE
Smart modeid pick
Loader
Shared lib. loader
Macros
Helper macros
Structs
All the structures
Logs
Library changelogs
Tools
Progs baked inside

Tools
amiscreen0.3
ansifilter0.1
auto2guide0.1
ccfilter0.1
chipset0.1
cputype0.1
device640.1
dupfilter0.1
expandlogs0.1
findsm0.2
fmpp0.1
fsmount0.1
genanim0.3
gensymtab0.1
hashlab0.4
ktm0.6
linkpoint0.1
logo2ilbm0.1
lpradio0.1
lzwpack0.1
minism0.2
mkheader0.1
modeidctrl0.1
mss0.6
multisum0.3
newterm0.2
numconv0.1
palhack1.1
patmat0.1
pfr0.2
playamitape0.1
qdbflags0.1
qfill0.1
resmodinit0.2
sadctrl0.2
saveassign0.1
setanim0.2
systime0.1
tokendump0.1
uassign0.1
viewlogo0.1
waitback0.4
wrapmount0.1



T
O
K
E
N
D
U
M
P

tokendump-0.1
--------------

---

This  program was written mainly  to create dictionaries from ordinary text
files. By default [0-9] [a-z] [A-Z] and Latin-1 characters will form tokens
but this can be changed by specifying DELIMCHARS.

---

NOTES:

[*]
Requires  68020(no FPU)+,  OS2.04(theoretically)+, 16+ kilo of free memory.

[*]
For  your convinience i have  added an  ISO 8859  character map, so you can
easily pick on  DELIMCHARS. On your right a table that contains  chars that
will form tokens.

     0 1 2 3 4 5 6 7 8 9 a b c d e f      0 1 2 3 4 5 6 7 8 9 a b c d e f
                                       
  0                                    0                                 
  1                                    1                                 
  2    ! " # $ % & ' ( ) * + , - . /   2                                 
  3  0 1 2 3 4 5 6 7 8 9 : ; < = > ?   3  0 1 2 3 4 5 6 7 8 9            
  4  @ A B C D E F G H I J K L M N O   4    A B C D E F G H I J K L M N O
  5  P Q R S T U V W X Y Z [ \ ] ^ _   5  P Q R S T U V W X Y Z          
  6  ` a b c d e f g h i j k l m n o   6    a b c d e f g h i j k l m n o
  7  p q r s t u v w x y z { | } ~    7  p q r s t u v w x y z          
  8                                    8                                 
  9                                    9                                 
  a    ¡ ¢ £ ¤ ¥ ¦ § ¨ © ª « ¬ ­ ® ¯   a                                 
  b  ° ± ² ³ ´ µ ¶ · ¸ ¹ º » ¼ ½ ¾ ¿   b                                 
  c  À Á Â Ã Ä Å Æ Ç È É Ê Ë Ì Í Î Ï   c  À Á Â Ã Ä Å Æ Ç È É Ê Ë Ì Í Î Ï
  d  Ð Ñ Ò Ó Ô Õ Ö × Ø Ù Ú Û Ü Ý Þ ß   d  Ð Ñ Ò Ó Ô Õ Ö × Ø Ù Ú Û Ü Ý Þ ß
  e  à á â ã ä å æ ç è é ê ë ì í î ï   e  à á â ã ä å æ ç è é ê ë ì í î ï
  f  ð ñ ò ó ô õ ö ÷ ø ù ú û ü ý þ ÿ   f  ð ñ ò ó ô õ ö ÷ ø ù ú û ü ý þ ÿ

[*]
After  tokenization pass the resulting file  to  'dupfilter'  to get rid of
possibly duplicated tokens.

---

HELP:

   > tokendump ?

     TEXTFILE/A,ND=NODIGITS/S,NL=NOLATIN/S,DELIMCHARS



   TEXTFILE/A     -  Text file to be wordified/tokenified. Max line size is
                     16 kilos.

   ND=NODIGITS/S  -  Tokens must not contain digits(0x30-0x39).

   NL=NOLATIN/S   -  Tokens must not contain Latin-1 characters(0xC0-0xFF).

   DELIMCHARS     -  Chars  that will  be  treated as a  delimiter. You can
                     specify them in  hexadecimal or decimal notation. Each
                     character  must be  comma separated. Passing  negative
                     value   means that this char must not be delimiter. By
                     default   0x01 - 0x1F,    0x21 - 0x2F,    0x3A - 0x40,
                     0x5B - 0x60   and   0x7B - 0xBF  will be translated to
                     0x20(delimiter).

---

USAGE:

   ; Tokenize text file allowing dashes and underscores
   tokendump <textfile> -0x2D,-0x5F >tokens.txt

---
megacz
    























No more fear cus pure HTML is here!
Copyright (C) 2013-2014 by Burnt Chip Dominators