forked from KolibriOS/kolibrios
tinypy: added re and random modules, not adapted for Kolibri yet.
git-svn-id: svn://kolibrios.org@2043 a494cfbc-eb01-0410-851d-a64ba20cac60
This commit is contained in:
52
programs/develop/tinypy/modules/random/init.c
Normal file
52
programs/develop/tinypy/modules/random/init.c
Normal file
@@ -0,0 +1,52 @@
|
|||||||
|
#include "random.c"
|
||||||
|
|
||||||
|
/*
|
||||||
|
* random_mod_init()
|
||||||
|
*
|
||||||
|
* random module initialization function
|
||||||
|
*/
|
||||||
|
void random_init(TP)
|
||||||
|
{
|
||||||
|
/*
|
||||||
|
* module dict for random
|
||||||
|
*/
|
||||||
|
tp_obj random_mod = tp_dict(tp);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bind functions to random module
|
||||||
|
*/
|
||||||
|
tp_set(tp, random_mod, tp_string("seed"), tp_fnc(tp, random_seed));
|
||||||
|
tp_set(tp, random_mod, tp_string("getstate"), tp_fnc(tp, random_getstate));
|
||||||
|
tp_set(tp, random_mod, tp_string("setstate"), tp_fnc(tp, random_setstate));
|
||||||
|
tp_set(tp, random_mod, tp_string("jumpahead"), tp_fnc(tp, random_jumpahead));
|
||||||
|
tp_set(tp, random_mod, tp_string("random"), tp_fnc(tp, random_random));
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bind usual distribution random variable generator
|
||||||
|
*/
|
||||||
|
tp_set(tp, random_mod, tp_string("uniform"), tp_fnc(tp, random_uniform));
|
||||||
|
tp_set(tp, random_mod, tp_string("normalvariate"), tp_fnc(tp, random_normalvariate));
|
||||||
|
tp_set(tp, random_mod, tp_string("lognormvariate"), tp_fnc(tp, random_lognormvariate));
|
||||||
|
tp_set(tp, random_mod, tp_string("expovariate"), tp_fnc(tp, random_expovariate));
|
||||||
|
tp_set(tp, random_mod, tp_string("vonmisesvariate"), tp_fnc(tp, random_vonmisesvariate));
|
||||||
|
tp_set(tp, random_mod, tp_string("gammavariate"), tp_fnc(tp, random_gammavariate));
|
||||||
|
tp_set(tp, random_mod, tp_string("betavariate"), tp_fnc(tp, random_betavariate));
|
||||||
|
tp_set(tp, random_mod, tp_string("paretovariate"), tp_fnc(tp, random_paretovariate));
|
||||||
|
tp_set(tp, random_mod, tp_string("weibullvariate"), tp_fnc(tp, random_weibullvariate));
|
||||||
|
tp_set(tp, random_mod, tp_string("randrange"), tp_fnc(tp, random_randrange));
|
||||||
|
tp_set(tp, random_mod, tp_string("randint"), tp_fnc(tp, random_randint));
|
||||||
|
tp_set(tp, random_mod, tp_string("choice"), tp_fnc(tp, random_choice));
|
||||||
|
tp_set(tp, random_mod, tp_string("shuffle"), tp_fnc(tp, random_shuffle));
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bind special attributes to random module
|
||||||
|
*/
|
||||||
|
tp_set(tp, random_mod, tp_string("__doc__"), tp_string("Random variable generators."));
|
||||||
|
tp_set(tp, random_mod, tp_string("__name__"), tp_string("random"));
|
||||||
|
tp_set(tp, random_mod, tp_string("__file__"), tp_string(__FILE__));
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bind random module to tinypy modules[]
|
||||||
|
*/
|
||||||
|
tp_set(tp, tp->modules, tp_string("random"), random_mod);
|
||||||
|
}
|
1107
programs/develop/tinypy/modules/random/random.c
Normal file
1107
programs/develop/tinypy/modules/random/random.c
Normal file
File diff suppressed because it is too large
Load Diff
176
programs/develop/tinypy/modules/random/tests.py
Normal file
176
programs/develop/tinypy/modules/random/tests.py
Normal file
@@ -0,0 +1,176 @@
|
|||||||
|
#!/usr/bin/env python
|
||||||
|
|
||||||
|
import random
|
||||||
|
#from math import log, exp, sqrt, pi
|
||||||
|
|
||||||
|
def test_seed_state():
|
||||||
|
"""test seed() and getstate()/setstate()
|
||||||
|
"""
|
||||||
|
# random ought to be able to deal with seeds in any form, of follows.
|
||||||
|
# following code shouldn't cause an exception.
|
||||||
|
random.seed()
|
||||||
|
random.seed(0)
|
||||||
|
random.seed(-1)
|
||||||
|
random.seed(0.1)
|
||||||
|
random.seed(-0.1)
|
||||||
|
random.seed("a")
|
||||||
|
random.seed("abc")
|
||||||
|
random.seed("abcd")
|
||||||
|
random.seed("fasdfasdfasdfadgaldhgldahlgahdlghadlgladh")
|
||||||
|
random.seed("lxhlh90yowhldshlgah;")
|
||||||
|
|
||||||
|
# state1 and state2 should be different for different seeds
|
||||||
|
random.seed(1)
|
||||||
|
state1 = random.getstate()
|
||||||
|
random.seed(2)
|
||||||
|
state2 = random.getstate()
|
||||||
|
rep = 0
|
||||||
|
for ind in range(len(state1)):
|
||||||
|
elem1 = state1[ind]
|
||||||
|
elem2 = state2[ind]
|
||||||
|
if (elem1 == elem2): rep += 1
|
||||||
|
if (rep > len(state1) / 2):
|
||||||
|
print("rep = ", rep, "len(state1) = ", len(state1))
|
||||||
|
raise "state1 and state2 should be different"
|
||||||
|
|
||||||
|
# for the same seeds, state1 and state2 should be the same
|
||||||
|
random.seed(100)
|
||||||
|
state1 = random.getstate()
|
||||||
|
random.seed(100)
|
||||||
|
state2 = random.getstate()
|
||||||
|
rep = 0
|
||||||
|
for ind in range(len(state1)):
|
||||||
|
elem1 = state1[ind]
|
||||||
|
elem2 = state2[ind]
|
||||||
|
if (elem1 == elem2): rep += 1
|
||||||
|
if (rep != len(state1)):
|
||||||
|
raise "state1 and state2 should be the same"
|
||||||
|
|
||||||
|
def test_jumpahead():
|
||||||
|
"""jumpahead will change the pseudo-number generator's internal state
|
||||||
|
"""
|
||||||
|
random.seed()
|
||||||
|
state1 = random.getstate()
|
||||||
|
random.jumpahead(20)
|
||||||
|
state2 = random.getstate()
|
||||||
|
rep = 0
|
||||||
|
for ind in range(len(state1)):
|
||||||
|
elem1 = state1[ind]
|
||||||
|
elem2 = state2[ind]
|
||||||
|
if (elem1 == elem2): rep += 1
|
||||||
|
if (rep > len(state1) / 2):
|
||||||
|
raise "state1 and state2 can't be the same"
|
||||||
|
|
||||||
|
def test_setstate():
|
||||||
|
"""
|
||||||
|
"""
|
||||||
|
random.seed()
|
||||||
|
oldState = random.getstate()
|
||||||
|
oldRandSeq = [random.random() for i in range(10)]
|
||||||
|
random.setstate(oldState)
|
||||||
|
newRandSeq = [random.random() for i in range(10)]
|
||||||
|
rep = 0
|
||||||
|
for ind in range(len(oldRandSeq)):
|
||||||
|
elem1 = oldRandSeq[ind]
|
||||||
|
elem2 = newRandSeq[ind]
|
||||||
|
if (elem1 == elem2): rep += 1
|
||||||
|
if (rep != len(oldRandSeq)):
|
||||||
|
raise "oldRandSeq and newRandSeq should be the same"
|
||||||
|
|
||||||
|
def test_random():
|
||||||
|
"""generate a random number list
|
||||||
|
"""
|
||||||
|
x = [random.random() for i in range(100)]
|
||||||
|
|
||||||
|
def test_distribution():
|
||||||
|
"""these lines are borrowed from python, they shouldn't
|
||||||
|
cause any exception.
|
||||||
|
"""
|
||||||
|
g = random
|
||||||
|
g.uniform(1,10)
|
||||||
|
g.paretovariate(1.0)
|
||||||
|
g.expovariate(1.0)
|
||||||
|
g.weibullvariate(1.0, 1.0)
|
||||||
|
g.normalvariate(0.0, 1.0)
|
||||||
|
g.lognormvariate(0.0, 1.0)
|
||||||
|
g.vonmisesvariate(0.0, 1.0)
|
||||||
|
g.gammavariate(0.01, 1.0)
|
||||||
|
g.gammavariate(1.0, 1.0)
|
||||||
|
g.gammavariate(200.0, 1.0)
|
||||||
|
g.betavariate(3.0, 3.0)
|
||||||
|
|
||||||
|
def test_randrange():
|
||||||
|
"""these input to randrange() shouldn't cause any exception.
|
||||||
|
"""
|
||||||
|
random.randrange(100000)
|
||||||
|
random.randrange(-100000)
|
||||||
|
random.randrange(0)
|
||||||
|
random.randrange(-10.2)
|
||||||
|
|
||||||
|
random.randrange(-10, 10)
|
||||||
|
random.randrange(2, 1000)
|
||||||
|
random.randrange(0, 1)
|
||||||
|
random.randrange(-1, 0)
|
||||||
|
|
||||||
|
random.randrange(10, 2000, 2)
|
||||||
|
random.randrange(-2000, 100, 5)
|
||||||
|
random.randrange(-1000.3, 1000.7, 2)
|
||||||
|
|
||||||
|
def test_randint():
|
||||||
|
"""for any valid pair (a, b), randint(a, b) should lay between [a, b]
|
||||||
|
"""
|
||||||
|
for i in range(1000):
|
||||||
|
r = random.randint(-10000, 10000)
|
||||||
|
if (-10000 <= r <= 10000): continue
|
||||||
|
else: raise "error: random.randint()"
|
||||||
|
|
||||||
|
def test_choice():
|
||||||
|
"""random.choice() should be able to deal with string, list.
|
||||||
|
"""
|
||||||
|
S = "abcdefg123*@#$%)("
|
||||||
|
L = [1, 2, 3, -1, 0.2, -0.1, -10000, "cyc"]
|
||||||
|
|
||||||
|
if random.choice(S) not in S:
|
||||||
|
raise "error: random.choice(S)"
|
||||||
|
|
||||||
|
if random.choice(L) not in L:
|
||||||
|
raise "error: random.choice(L)"
|
||||||
|
|
||||||
|
def test_shuffle():
|
||||||
|
"""test random.shuffle() on list. since string is not writable in-place,
|
||||||
|
random.shuffle() can not be applied on string.
|
||||||
|
Note: to copy items from a list to a new list, must use syntax like:
|
||||||
|
newList = oldList[:]
|
||||||
|
if use syntax like: newList = oldList, newList is just an alias of oldList.
|
||||||
|
"""
|
||||||
|
oldL = [1, 2, 3, -1, 0.2, -0.1, -10000, "cyc"]
|
||||||
|
newL = oldL[:]
|
||||||
|
|
||||||
|
random.shuffle(newL)
|
||||||
|
|
||||||
|
rep = 0
|
||||||
|
for ind in range(len(oldL)):
|
||||||
|
elem1 = oldL[ind]
|
||||||
|
elem2 = newL[ind]
|
||||||
|
if (elem1 == elem2): rep += 1
|
||||||
|
if (rep > len(oldL) / 2):
|
||||||
|
raise "oldL and newL shouldn't be the same"
|
||||||
|
|
||||||
|
def test_53_bits_per_float():
|
||||||
|
pass
|
||||||
|
|
||||||
|
def main():
|
||||||
|
test_seed_state()
|
||||||
|
test_jumpahead()
|
||||||
|
test_setstate()
|
||||||
|
test_random()
|
||||||
|
test_distribution()
|
||||||
|
test_randrange()
|
||||||
|
test_randint()
|
||||||
|
test_choice()
|
||||||
|
test_shuffle()
|
||||||
|
test_53_bits_per_float()
|
||||||
|
print("#OK")
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
main()
|
710
programs/develop/tinypy/modules/re/init.c
Normal file
710
programs/develop/tinypy/modules/re/init.c
Normal file
@@ -0,0 +1,710 @@
|
|||||||
|
/*
|
||||||
|
* regular expression module
|
||||||
|
*
|
||||||
|
* Important Note: do not support group name index
|
||||||
|
*
|
||||||
|
* $Id$
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <stdio.h>
|
||||||
|
#include <assert.h>
|
||||||
|
#include "regexpr.c"
|
||||||
|
|
||||||
|
/* tinypy API to be use in this unit */
|
||||||
|
extern tp_obj tp_data(TP,int magic,void *v);
|
||||||
|
extern tp_obj tp_object_new(TP);
|
||||||
|
extern tp_obj tp_object(TP);
|
||||||
|
extern tp_obj tp_method(TP,tp_obj self,tp_obj v(TP));
|
||||||
|
extern tp_obj tp_string_copy(TP, const char *s, int n);
|
||||||
|
extern tp_obj tp_list(TP);
|
||||||
|
extern tp_obj tp_copy(TP);
|
||||||
|
|
||||||
|
/* last error message */
|
||||||
|
static const char * LastError = NULL;
|
||||||
|
|
||||||
|
/* lower level regex object */
|
||||||
|
typedef struct {
|
||||||
|
struct re_pattern_buffer re_patbuf; /* The compiled expression */
|
||||||
|
struct re_registers re_regs; /* The registers from the last match */
|
||||||
|
char re_fastmap[256]; /* Storage for fastmap */
|
||||||
|
unsigned char *re_translate; /* String object for translate table */
|
||||||
|
unsigned char *re_lastok; /* String object last matched/searched */
|
||||||
|
|
||||||
|
/* supplementary */
|
||||||
|
int re_errno; /* error num */
|
||||||
|
int re_syntax; /* syntax */
|
||||||
|
} regexobject;
|
||||||
|
|
||||||
|
/* local declarations */
|
||||||
|
static regexobject* getre(TP, tp_obj rmobj);
|
||||||
|
static tp_obj match_obj_group(TP);
|
||||||
|
static tp_obj match_obj_groups(TP);
|
||||||
|
static tp_obj match_obj_start(TP);
|
||||||
|
static tp_obj match_obj_end(TP);
|
||||||
|
static tp_obj match_obj_span(TP);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* helper function: return lower level regex object
|
||||||
|
* rmobj - regex or match object
|
||||||
|
*/
|
||||||
|
static regexobject * getre(TP, tp_obj rmobj)
|
||||||
|
{
|
||||||
|
tp_obj reobj_data = tp_get(tp, rmobj, tp_string("__data__"));
|
||||||
|
regexobject *re = NULL;
|
||||||
|
|
||||||
|
/* validate magic */
|
||||||
|
if (reobj_data.data.magic != sizeof(regexobject)) {
|
||||||
|
LastError = "broken regex object";
|
||||||
|
return (NULL);
|
||||||
|
}
|
||||||
|
re = (regexobject*)reobj_data.data.val;
|
||||||
|
assert(re);
|
||||||
|
|
||||||
|
return (re);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* derive match object from regex object
|
||||||
|
*/
|
||||||
|
static tp_obj match_object(TP, tp_obj reobj)
|
||||||
|
{
|
||||||
|
tp_obj mo = tp_object(tp); /* match object */
|
||||||
|
tp_obj redata; /* regex object data */
|
||||||
|
tp_obj madata; /* match object data */
|
||||||
|
regexobject *re = NULL; /* lower level regex object */
|
||||||
|
|
||||||
|
redata = tp_get(tp, reobj, tp_string("__data__"));
|
||||||
|
re = (regexobject *)redata.data.val;
|
||||||
|
assert(re);
|
||||||
|
madata = tp_data(tp, (int)sizeof(regexobject), re);
|
||||||
|
|
||||||
|
tp_set(tp, mo, tp_string("group"), tp_method(tp, mo, match_obj_group));
|
||||||
|
tp_set(tp, mo, tp_string("groups"), tp_method(tp, mo, match_obj_groups));
|
||||||
|
tp_set(tp, mo, tp_string("start"), tp_method(tp, mo, match_obj_start));
|
||||||
|
tp_set(tp, mo, tp_string("end"), tp_method(tp, mo, match_obj_end));
|
||||||
|
tp_set(tp, mo, tp_string("span"), tp_method(tp, mo, match_obj_span));
|
||||||
|
tp_set(tp, mo, tp_string("__data__"), madata);
|
||||||
|
|
||||||
|
return (mo);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* FUNC: regexobj.search(str[,pos=0])
|
||||||
|
* self - regex object
|
||||||
|
* str - string to be searched
|
||||||
|
* pos - optional starting offset
|
||||||
|
*
|
||||||
|
* RETURN:
|
||||||
|
* match object - when matched
|
||||||
|
* None - not matched
|
||||||
|
*/
|
||||||
|
static tp_obj regex_obj_search(TP)
|
||||||
|
{
|
||||||
|
tp_obj self = TP_OBJ(); /* regex object */
|
||||||
|
tp_obj str = TP_STR();
|
||||||
|
tp_obj pos = TP_DEFAULT(tp_number(0));
|
||||||
|
tp_obj maobj; /* match object */
|
||||||
|
regexobject *re = NULL;
|
||||||
|
int r = -2; /* -2 indicate exception */
|
||||||
|
int range;
|
||||||
|
|
||||||
|
if (pos.number.val < 0 || pos.number.val > str.string.len) {
|
||||||
|
LastError = "search offset out of range";
|
||||||
|
goto exception;
|
||||||
|
}
|
||||||
|
range = str.string.len - pos.number.val;
|
||||||
|
|
||||||
|
re = getre(tp, self);
|
||||||
|
re->re_lastok = NULL;
|
||||||
|
r = re_search(&re->re_patbuf, (unsigned char *)str.string.val,
|
||||||
|
str.string.len, pos.number.val, range, &re->re_regs);
|
||||||
|
|
||||||
|
/* cannot match pattern */
|
||||||
|
if (r == -1)
|
||||||
|
goto notfind;
|
||||||
|
|
||||||
|
/* error occurred */
|
||||||
|
if (r == -2)
|
||||||
|
goto exception;
|
||||||
|
|
||||||
|
/* matched */
|
||||||
|
re->re_lastok = (unsigned char *)str.string.val;
|
||||||
|
|
||||||
|
/* match obj */
|
||||||
|
maobj = match_object(tp, self);
|
||||||
|
|
||||||
|
return (maobj);
|
||||||
|
|
||||||
|
notfind:
|
||||||
|
re->re_lastok = NULL;
|
||||||
|
return (tp_None);
|
||||||
|
exception:
|
||||||
|
re->re_lastok = NULL;
|
||||||
|
tp_raise(tp_None, tp_string("regex search error"));
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* FUNC: regexobj.match(str[,pos=0])
|
||||||
|
* self - regex object
|
||||||
|
* str - string to be matched
|
||||||
|
* pos - optional starting position
|
||||||
|
*
|
||||||
|
* RETURN:
|
||||||
|
* match object - when matched
|
||||||
|
* None - not matched
|
||||||
|
*/
|
||||||
|
static tp_obj regex_obj_match(TP)
|
||||||
|
{
|
||||||
|
tp_obj self = TP_OBJ(); /* regex object */
|
||||||
|
tp_obj str = TP_STR();
|
||||||
|
tp_obj pos = TP_DEFAULT(tp_number(0));
|
||||||
|
tp_obj maobj; /* match object */
|
||||||
|
regexobject *re = NULL;
|
||||||
|
int r = -2; /* -2 indicate exception */
|
||||||
|
|
||||||
|
re = getre(tp, self);
|
||||||
|
re->re_lastok = NULL;
|
||||||
|
r = re_match(&re->re_patbuf, (unsigned char *)str.string.val,
|
||||||
|
str.string.len, pos.number.val, &re->re_regs);
|
||||||
|
|
||||||
|
/* cannot match pattern */
|
||||||
|
if (r == -1)
|
||||||
|
goto nomatch;
|
||||||
|
|
||||||
|
/* error occurred */
|
||||||
|
if (r == -2)
|
||||||
|
goto exception;
|
||||||
|
|
||||||
|
/* matched */
|
||||||
|
re->re_lastok = (unsigned char *)str.string.val;
|
||||||
|
|
||||||
|
/* match obj */
|
||||||
|
maobj = match_object(tp, self);
|
||||||
|
|
||||||
|
return (maobj);
|
||||||
|
|
||||||
|
nomatch:
|
||||||
|
re->re_lastok = NULL;
|
||||||
|
return (tp_None);
|
||||||
|
exception:
|
||||||
|
re->re_lastok = NULL;
|
||||||
|
tp_raise(tp_None, tp_string("regex match error"));
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* regex object split()
|
||||||
|
* self - regex object
|
||||||
|
* restr - regex string
|
||||||
|
* maxsplit - max split field, default 0, mean no limit
|
||||||
|
*/
|
||||||
|
static tp_obj regex_obj_split(TP)
|
||||||
|
{
|
||||||
|
tp_obj self = TP_OBJ(); /* regex object */
|
||||||
|
tp_obj restr = TP_OBJ(); /* string */
|
||||||
|
tp_obj maxsplit = TP_DEFAULT(tp_number(0));
|
||||||
|
tp_obj maobj; /* match object */
|
||||||
|
regexobject *re = NULL; /* lower level regex object */
|
||||||
|
tp_obj result = tp_list(tp);
|
||||||
|
tp_obj grpstr; /* group string */
|
||||||
|
int slen; /* string length */
|
||||||
|
int srchloc; /* search location */
|
||||||
|
|
||||||
|
/* maxsplit == 0 means no limit */
|
||||||
|
if ((int)maxsplit.number.val == 0)
|
||||||
|
maxsplit.number.val = RE_NREGS;
|
||||||
|
assert(maxsplit.number.val > 0);
|
||||||
|
|
||||||
|
srchloc = 0;
|
||||||
|
slen = strlen((char *)restr.string.val);
|
||||||
|
|
||||||
|
do {
|
||||||
|
/* generate a temp match object */
|
||||||
|
tp_params_v(tp, 3, self, restr, tp_number(srchloc));
|
||||||
|
maobj = regex_obj_search(tp);
|
||||||
|
if (!tp_bool(tp, maobj))
|
||||||
|
break;
|
||||||
|
|
||||||
|
re = getre(tp, maobj);
|
||||||
|
if (re->re_lastok == NULL) {
|
||||||
|
tp_raise(tp_None, tp_string("no match for split()"));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* extract fields */
|
||||||
|
if ((int)maxsplit.number.val > 0) {
|
||||||
|
int start = re->re_regs.start[0];
|
||||||
|
int end = re->re_regs.end[0];
|
||||||
|
/*printf("%s:start(%d),end(%d)\n", __func__, start, end);*/
|
||||||
|
if (start < 0 || end < 0)
|
||||||
|
break;
|
||||||
|
|
||||||
|
grpstr = tp_string_copy(tp,
|
||||||
|
(const char *)re->re_lastok + srchloc, start - srchloc);
|
||||||
|
|
||||||
|
if (tp_bool(tp, grpstr)) {
|
||||||
|
tp_set(tp, result, tp_None, grpstr);
|
||||||
|
maxsplit.number.val--;
|
||||||
|
}
|
||||||
|
|
||||||
|
srchloc = end;
|
||||||
|
}
|
||||||
|
} while (srchloc < slen && (int)maxsplit.number.val > 0);
|
||||||
|
|
||||||
|
/* collect remaining string, if necessary */
|
||||||
|
if (srchloc < slen) {
|
||||||
|
grpstr = tp_string_copy(tp,
|
||||||
|
(const char *)restr.string.val + srchloc, slen - srchloc);
|
||||||
|
if (tp_bool(tp, grpstr))
|
||||||
|
tp_set(tp, result, tp_None, grpstr);
|
||||||
|
}
|
||||||
|
|
||||||
|
return (result);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* regex object findall()
|
||||||
|
* self - regex object
|
||||||
|
* restr - regex string
|
||||||
|
* pos - starting position, default 0
|
||||||
|
*/
|
||||||
|
static tp_obj regex_obj_findall(TP)
|
||||||
|
{
|
||||||
|
tp_obj self = TP_OBJ(); /* regex object */
|
||||||
|
tp_obj restr = TP_OBJ(); /* string */
|
||||||
|
tp_obj pos = TP_DEFAULT(tp_number(0));
|
||||||
|
tp_obj maobj; /* match object */
|
||||||
|
regexobject *re = NULL; /* lower level regex object */
|
||||||
|
tp_obj result = tp_list(tp);
|
||||||
|
tp_obj grpstr; /* group string */
|
||||||
|
int slen; /* string length */
|
||||||
|
int srchloc; /* search location */
|
||||||
|
|
||||||
|
srchloc = (int)pos.number.val;
|
||||||
|
slen = strlen((char *)restr.string.val);
|
||||||
|
if (srchloc < 0 || srchloc >= slen)
|
||||||
|
tp_raise(tp_None, tp_string("starting position out of range"));
|
||||||
|
|
||||||
|
do {
|
||||||
|
/* generate a temp match object */
|
||||||
|
tp_params_v(tp, 3, self, restr, tp_number(srchloc));
|
||||||
|
maobj = regex_obj_search(tp);
|
||||||
|
if (!tp_bool(tp, maobj))
|
||||||
|
break;
|
||||||
|
|
||||||
|
re = getre(tp, maobj);
|
||||||
|
if (re->re_lastok == NULL) {
|
||||||
|
tp_raise(tp_None, tp_string("no match for findall()"));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* extract fields */
|
||||||
|
if (srchloc < slen) {
|
||||||
|
int start = re->re_regs.start[0];
|
||||||
|
int end = re->re_regs.end[0];
|
||||||
|
/*printf("%s:start(%d),end(%d)\n", __func__, start, end);*/
|
||||||
|
if (start < 0 || end < 0)
|
||||||
|
break;
|
||||||
|
|
||||||
|
grpstr = tp_string_copy(tp,
|
||||||
|
(const char *)re->re_lastok + start, end - start);
|
||||||
|
|
||||||
|
if (tp_bool(tp, grpstr)) {
|
||||||
|
tp_set(tp, result, tp_None, grpstr);
|
||||||
|
}
|
||||||
|
|
||||||
|
srchloc = end;
|
||||||
|
}
|
||||||
|
} while (srchloc < slen);
|
||||||
|
|
||||||
|
return (result);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* FUNC: matchobj.group([group1, ...])
|
||||||
|
* self - match object
|
||||||
|
* args - optional group indices, default 0
|
||||||
|
*
|
||||||
|
* return specified group.
|
||||||
|
*/
|
||||||
|
static tp_obj match_obj_group(TP)
|
||||||
|
{
|
||||||
|
tp_obj self = TP_OBJ(); /* match object */
|
||||||
|
tp_obj grpidx; /* a group index */
|
||||||
|
regexobject *re = NULL;
|
||||||
|
int indices[RE_NREGS];
|
||||||
|
int start;
|
||||||
|
int end;
|
||||||
|
int i;
|
||||||
|
int single = 0; /* single group index? */
|
||||||
|
tp_obj result;
|
||||||
|
|
||||||
|
/* get lower level regex object representation */
|
||||||
|
re = getre(tp, self);
|
||||||
|
if (re->re_lastok == NULL)
|
||||||
|
tp_raise(tp_None,
|
||||||
|
tp_string("group() only valid after successful match/search"));
|
||||||
|
|
||||||
|
for (i = 0; i < RE_NREGS; i++)
|
||||||
|
indices[i] = -1;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* if no group index provided, supply default group index 0; else
|
||||||
|
* fill in indices[] with provided group index list.
|
||||||
|
*/
|
||||||
|
if (tp->params.list.val->len == 0) {
|
||||||
|
indices[0] = 0;
|
||||||
|
single = 1;
|
||||||
|
} else if (tp->params.list.val->len == 1) {
|
||||||
|
indices[0] = (int)TP_NUM();
|
||||||
|
single = 1;
|
||||||
|
} else {
|
||||||
|
i = 0;
|
||||||
|
TP_LOOP(grpidx)
|
||||||
|
if (grpidx.number.val < 0 || grpidx.number.val > RE_NREGS)
|
||||||
|
tp_raise(tp_None, tp_string("group() grpidx out of range"));
|
||||||
|
indices[i++] = (int)grpidx.number.val;
|
||||||
|
TP_END
|
||||||
|
}
|
||||||
|
|
||||||
|
/* generate result string list */
|
||||||
|
result = tp_list(tp);
|
||||||
|
for (i = 0; i < RE_NREGS && indices[i] >= 0; i++) {
|
||||||
|
tp_obj grpstr;
|
||||||
|
start = re->re_regs.start[indices[i]];
|
||||||
|
end = re->re_regs.end[indices[i]];
|
||||||
|
if (start < 0 || end < 0) {
|
||||||
|
grpstr = tp_None;
|
||||||
|
} else {
|
||||||
|
grpstr = tp_string_copy(tp, (const char *)re->re_lastok + start,
|
||||||
|
end - start);
|
||||||
|
}
|
||||||
|
tp_set(tp, result, tp_None, grpstr);
|
||||||
|
}
|
||||||
|
return (single ? tp_get(tp, result, tp_number(0)) : result);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* FUNC: matchobj.groups()
|
||||||
|
* self - match object.
|
||||||
|
* return all groups.
|
||||||
|
* Note: CPython allow a 'default' argument, but we disallow it.
|
||||||
|
*/
|
||||||
|
static tp_obj match_obj_groups(TP)
|
||||||
|
{
|
||||||
|
tp_obj self = TP_OBJ(); /* match object */
|
||||||
|
regexobject *re = NULL;
|
||||||
|
int start;
|
||||||
|
int end;
|
||||||
|
int i;
|
||||||
|
tp_obj result = tp_list(tp);
|
||||||
|
|
||||||
|
re = getre(tp, self);
|
||||||
|
if (re->re_lastok == NULL) {
|
||||||
|
tp_raise(tp_None,
|
||||||
|
tp_string("groups() only valid after successful match/search"));
|
||||||
|
}
|
||||||
|
|
||||||
|
for (i = 1; i < RE_NREGS; i++) {
|
||||||
|
start = re->re_regs.start[i];
|
||||||
|
end = re->re_regs.end[i];
|
||||||
|
if (start < 0 || end < 0)
|
||||||
|
break;
|
||||||
|
|
||||||
|
tp_obj grpstr = tp_string_copy(tp,
|
||||||
|
(const char *)re->re_lastok + start, end - start);
|
||||||
|
|
||||||
|
if (tp_bool(tp, grpstr))
|
||||||
|
tp_set(tp, result, tp_None, grpstr);
|
||||||
|
}
|
||||||
|
|
||||||
|
return (result);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* FUNC: matchobj.start([group])
|
||||||
|
* self - match object
|
||||||
|
* group - group index
|
||||||
|
* return starting position of matched 'group' substring.
|
||||||
|
*/
|
||||||
|
static tp_obj match_obj_start(TP)
|
||||||
|
{
|
||||||
|
tp_obj self = TP_OBJ(); /* match object */
|
||||||
|
tp_obj group = TP_DEFAULT(tp_number(0)); /* group */
|
||||||
|
regexobject *re = NULL;
|
||||||
|
int start;
|
||||||
|
|
||||||
|
re = getre(tp, self);
|
||||||
|
if (re->re_lastok == NULL) {
|
||||||
|
tp_raise(tp_None,
|
||||||
|
tp_string("start() only valid after successful match/search"));
|
||||||
|
}
|
||||||
|
|
||||||
|
if (group.number.val < 0 || group.number.val > RE_NREGS)
|
||||||
|
tp_raise(tp_None, tp_string("IndexError: group index out of range"));
|
||||||
|
|
||||||
|
start = re->re_regs.start[(int)group.number.val];
|
||||||
|
|
||||||
|
return (tp_number(start));
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* FUNC: matchobj.end([group])
|
||||||
|
* self - match object
|
||||||
|
* group - group index
|
||||||
|
* return ending position of matched 'group' substring.
|
||||||
|
*/
|
||||||
|
static tp_obj match_obj_end(TP)
|
||||||
|
{
|
||||||
|
tp_obj self = TP_OBJ(); /* match object */
|
||||||
|
tp_obj group = TP_DEFAULT(tp_number(0)); /* group */
|
||||||
|
regexobject *re = NULL;
|
||||||
|
int end;
|
||||||
|
|
||||||
|
re = getre(tp, self);
|
||||||
|
if (re->re_lastok == NULL) {
|
||||||
|
tp_raise(tp_None,
|
||||||
|
tp_string("end() only valid after successful match/search"));
|
||||||
|
}
|
||||||
|
|
||||||
|
if (group.number.val < 0 || group.number.val > RE_NREGS)
|
||||||
|
tp_raise(tp_None, tp_string("IndexError: group index out of range"));
|
||||||
|
|
||||||
|
end = re->re_regs.end[(int)group.number.val];
|
||||||
|
|
||||||
|
return (tp_number(end));
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* FUNC: matchobj.span([group])
|
||||||
|
* self - match object
|
||||||
|
* group - group index
|
||||||
|
* return [start,end] position pair of matched 'group' substring.
|
||||||
|
*/
|
||||||
|
static tp_obj match_obj_span(TP)
|
||||||
|
{
|
||||||
|
tp_obj self = TP_OBJ(); /* match object */
|
||||||
|
tp_obj group = TP_DEFAULT(tp_number(0)); /* group */
|
||||||
|
regexobject *re = NULL;
|
||||||
|
int start;
|
||||||
|
int end;
|
||||||
|
tp_obj result;
|
||||||
|
|
||||||
|
re = getre(tp, self);
|
||||||
|
if (re->re_lastok == NULL) {
|
||||||
|
tp_raise(tp_None,
|
||||||
|
tp_string("span() only valid after successful match/search"));
|
||||||
|
}
|
||||||
|
|
||||||
|
if (group.number.val < 0 || group.number.val > RE_NREGS)
|
||||||
|
tp_raise(tp_None, tp_string("IndexError: group index out of range"));
|
||||||
|
|
||||||
|
start = re->re_regs.start[(int)group.number.val];
|
||||||
|
end = re->re_regs.end[(int)group.number.val];
|
||||||
|
|
||||||
|
result = tp_list(tp);
|
||||||
|
tp_set(tp, result, tp_None, tp_number(start));
|
||||||
|
tp_set(tp, result, tp_None, tp_number(end));
|
||||||
|
|
||||||
|
return (result);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* compile out a re object
|
||||||
|
* repat - regex pattern
|
||||||
|
* resyn - regex syntax
|
||||||
|
*/
|
||||||
|
static tp_obj regex_compile(TP)
|
||||||
|
{
|
||||||
|
char *error = NULL;
|
||||||
|
char const *pat = NULL;
|
||||||
|
int size = 0;
|
||||||
|
tp_obj reobj_data;
|
||||||
|
tp_obj repat = TP_TYPE(TP_STRING); /* pattern */
|
||||||
|
tp_obj resyn = TP_DEFAULT(tp_number(RE_SYNTAX_EMACS)); /* syntax */
|
||||||
|
tp_obj reobj; /* regex object */
|
||||||
|
regexobject *re;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* create regex object, its parent is builtin 'object'
|
||||||
|
*/
|
||||||
|
reobj = tp_object(tp);
|
||||||
|
|
||||||
|
re = (regexobject *)malloc(sizeof(regexobject));
|
||||||
|
if (!re) {
|
||||||
|
error = "malloc lower level regex object failed";
|
||||||
|
goto finally;
|
||||||
|
}
|
||||||
|
|
||||||
|
re->re_patbuf.buffer = NULL;
|
||||||
|
re->re_patbuf.allocated = 0;
|
||||||
|
re->re_patbuf.fastmap = (unsigned char *)re->re_fastmap;
|
||||||
|
re->re_patbuf.translate = NULL;
|
||||||
|
re->re_translate = NULL;
|
||||||
|
re->re_lastok = NULL;
|
||||||
|
|
||||||
|
re->re_errno = 0;
|
||||||
|
re->re_syntax = (int)resyn.number.val;
|
||||||
|
|
||||||
|
pat = repat.string.val;
|
||||||
|
size = repat.string.len;
|
||||||
|
error = re_compile_pattern((unsigned char *)pat, size, &re->re_patbuf);
|
||||||
|
if (error != NULL) {
|
||||||
|
LastError = error;
|
||||||
|
goto finally;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* regexobject's size as magic */
|
||||||
|
reobj_data = tp_data(tp, (int)sizeof(regexobject), re);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bind to regex object
|
||||||
|
*/
|
||||||
|
tp_set(tp, reobj, tp_string("search"),
|
||||||
|
tp_method(tp, reobj, regex_obj_search));
|
||||||
|
tp_set(tp, reobj, tp_string("match"),
|
||||||
|
tp_method(tp, reobj, regex_obj_match));
|
||||||
|
tp_set(tp, reobj, tp_string("split"),
|
||||||
|
tp_method(tp, reobj, regex_obj_split));
|
||||||
|
tp_set(tp, reobj, tp_string("findall"),
|
||||||
|
tp_method(tp, reobj, regex_obj_findall));
|
||||||
|
tp_set(tp, reobj, tp_string("__data__"), reobj_data);
|
||||||
|
|
||||||
|
tp_set(tp, reobj, tp_string("__name__"),
|
||||||
|
tp_string("regular expression object"));
|
||||||
|
tp_set(tp, reobj, tp_string("__doc__"), tp_string(
|
||||||
|
"regular expression object, support methods:\n"
|
||||||
|
"search(str[,pos=0])-search 'str' from 'pos'\n"
|
||||||
|
"match(str[,pos=0]) -match 'str' from 'pos'\n"
|
||||||
|
));
|
||||||
|
|
||||||
|
return (reobj);
|
||||||
|
|
||||||
|
finally:
|
||||||
|
tp_raise(tp_None, tp_string(error));
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* module level search()
|
||||||
|
*/
|
||||||
|
static tp_obj regex_search(TP)
|
||||||
|
{
|
||||||
|
tp_obj repat = TP_OBJ(); /* pattern */
|
||||||
|
tp_obj restr = TP_OBJ(); /* string */
|
||||||
|
tp_obj resyn = TP_DEFAULT(tp_number(RE_SYNTAX_EMACS));
|
||||||
|
tp_obj reobj; /* regex object */
|
||||||
|
tp_obj maobj; /* match object */
|
||||||
|
|
||||||
|
/* compile out regex object */
|
||||||
|
tp_params_v(tp, 2, repat, resyn);
|
||||||
|
reobj = regex_compile(tp);
|
||||||
|
|
||||||
|
/* call r.search() */
|
||||||
|
tp_params_v(tp, 3, reobj, restr, tp_number(0));
|
||||||
|
maobj = regex_obj_search(tp);
|
||||||
|
|
||||||
|
return (maobj);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* module level match()
|
||||||
|
*/
|
||||||
|
static tp_obj regex_match(TP)
|
||||||
|
{
|
||||||
|
tp_obj repat = TP_OBJ(); /* pattern */
|
||||||
|
tp_obj restr = TP_OBJ(); /* string */
|
||||||
|
tp_obj resyn = TP_DEFAULT(tp_number(RE_SYNTAX_EMACS));
|
||||||
|
tp_obj reobj; /* regex object */
|
||||||
|
tp_obj maobj; /* match object */
|
||||||
|
|
||||||
|
/* compile out regex object */
|
||||||
|
tp_params_v(tp, 2, repat, resyn);
|
||||||
|
reobj = regex_compile(tp);
|
||||||
|
|
||||||
|
/* call r.search() */
|
||||||
|
tp_params_v(tp, 3, reobj, restr, tp_number(0));
|
||||||
|
maobj = regex_obj_match(tp);
|
||||||
|
|
||||||
|
return (maobj);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* module level split()
|
||||||
|
* repat - regex pattern
|
||||||
|
* restr - regex string
|
||||||
|
* maxsplit - max split field, default 0, mean no limit
|
||||||
|
*/
|
||||||
|
static tp_obj regex_split(TP)
|
||||||
|
{
|
||||||
|
tp_obj repat = TP_OBJ(); /* pattern */
|
||||||
|
tp_obj restr = TP_OBJ(); /* string */
|
||||||
|
tp_obj maxsplit = TP_DEFAULT(tp_number(0));
|
||||||
|
tp_obj reobj; /* regex object */
|
||||||
|
|
||||||
|
/* generate a temp regex object */
|
||||||
|
tp_params_v(tp, 2, repat, tp_number(RE_SYNTAX_EMACS));
|
||||||
|
reobj = regex_compile(tp);
|
||||||
|
|
||||||
|
tp_params_v(tp, 3, reobj, restr, maxsplit);
|
||||||
|
return regex_obj_split(tp);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* module level findall()
|
||||||
|
* repat - regex pattern
|
||||||
|
* restr - regex string
|
||||||
|
* resyn - regex syntax, optional, default RE_SYNTAX_EMAC
|
||||||
|
*/
|
||||||
|
static tp_obj regex_findall(TP)
|
||||||
|
{
|
||||||
|
tp_obj repat = TP_OBJ(); /* pattern */
|
||||||
|
tp_obj restr = TP_OBJ(); /* string */
|
||||||
|
tp_obj resyn = TP_DEFAULT(tp_number(RE_SYNTAX_EMACS));
|
||||||
|
tp_obj reobj; /* regex object */
|
||||||
|
|
||||||
|
/* generate a temp regex object */
|
||||||
|
tp_params_v(tp, 2, repat, resyn);
|
||||||
|
reobj = regex_compile(tp);
|
||||||
|
|
||||||
|
tp_params_v(tp, 2, reobj, restr);
|
||||||
|
return regex_obj_findall(tp);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/*
|
||||||
|
* re mod can only support 'set_syntax', 'get_syntax', and 'compile' functions,
|
||||||
|
* 'compile' function will return a 'reobj', and this 'reobj' will support
|
||||||
|
* methods 'search', 'match', 'group', 'groupall', el al.
|
||||||
|
*/
|
||||||
|
void re_init(TP)
|
||||||
|
{
|
||||||
|
/*
|
||||||
|
* module dict for re
|
||||||
|
*/
|
||||||
|
tp_obj re_mod = tp_dict(tp);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bind to re module
|
||||||
|
*/
|
||||||
|
tp_set(tp, re_mod, tp_string("compile"), tp_fnc(tp, regex_compile));
|
||||||
|
tp_set(tp, re_mod, tp_string("search"), tp_fnc(tp, regex_search));
|
||||||
|
tp_set(tp, re_mod, tp_string("match"), tp_fnc(tp, regex_match));
|
||||||
|
tp_set(tp, re_mod, tp_string("split"), tp_fnc(tp, regex_split));
|
||||||
|
tp_set(tp, re_mod, tp_string("findall"), tp_fnc(tp, regex_findall));
|
||||||
|
tp_set(tp, re_mod, tp_string("AWK_SYNTAX"), tp_number(RE_SYNTAX_AWK));
|
||||||
|
tp_set(tp, re_mod, tp_string("EGREP_SYNTAX"), tp_number(RE_SYNTAX_EGREP));
|
||||||
|
tp_set(tp, re_mod, tp_string("GREP_SYNTAX"), tp_number(RE_SYNTAX_GREP));
|
||||||
|
tp_set(tp, re_mod, tp_string("EMACS_SYNTAX"), tp_number(RE_SYNTAX_EMACS));
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bind special attibutes to re module
|
||||||
|
*/
|
||||||
|
tp_set(tp, re_mod, tp_string("__name__"),
|
||||||
|
tp_string("regular expression module"));
|
||||||
|
tp_set(tp, re_mod, tp_string("__file__"), tp_string(__FILE__));
|
||||||
|
tp_set(tp, re_mod, tp_string("__doc__"),
|
||||||
|
tp_string("simple regular express implementation"));
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bind regex module to tinypy modules[]
|
||||||
|
*/
|
||||||
|
tp_set(tp, tp->modules, tp_string("re"), re_mod);
|
||||||
|
}
|
||||||
|
|
2124
programs/develop/tinypy/modules/re/regexpr.c
Normal file
2124
programs/develop/tinypy/modules/re/regexpr.c
Normal file
File diff suppressed because it is too large
Load Diff
160
programs/develop/tinypy/modules/re/regexpr.h
Normal file
160
programs/develop/tinypy/modules/re/regexpr.h
Normal file
@@ -0,0 +1,160 @@
|
|||||||
|
/*
|
||||||
|
* -*- mode: c-mode; c-file-style: python -*-
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef Py_REGEXPR_H
|
||||||
|
#define Py_REGEXPR_H
|
||||||
|
#ifdef __cplusplus
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*
|
||||||
|
* regexpr.h
|
||||||
|
*
|
||||||
|
* Author: Tatu Ylonen <ylo@ngs.fi>
|
||||||
|
*
|
||||||
|
* Copyright (c) 1991 Tatu Ylonen, Espoo, Finland
|
||||||
|
*
|
||||||
|
* Permission to use, copy, modify, distribute, and sell this software
|
||||||
|
* and its documentation for any purpose is hereby granted without fee,
|
||||||
|
* provided that the above copyright notice appear in all copies. This
|
||||||
|
* software is provided "as is" without express or implied warranty.
|
||||||
|
*
|
||||||
|
* Created: Thu Sep 26 17:15:36 1991 ylo
|
||||||
|
* Last modified: Mon Nov 4 15:49:46 1991 ylo
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* $Id$ */
|
||||||
|
|
||||||
|
#ifndef REGEXPR_H
|
||||||
|
#define REGEXPR_H
|
||||||
|
|
||||||
|
#define RE_NREGS 100 /* number of registers available */
|
||||||
|
|
||||||
|
typedef struct re_pattern_buffer
|
||||||
|
{
|
||||||
|
unsigned char *buffer; /* compiled pattern */
|
||||||
|
int allocated; /* allocated size of compiled pattern */
|
||||||
|
int used; /* actual length of compiled pattern */
|
||||||
|
unsigned char *fastmap; /* fastmap[ch] is true if ch can start pattern */
|
||||||
|
unsigned char *translate; /* translation to apply during compilation/matching */
|
||||||
|
unsigned char fastmap_accurate; /* true if fastmap is valid */
|
||||||
|
unsigned char can_be_null; /* true if can match empty string */
|
||||||
|
unsigned char uses_registers; /* registers are used and need to be initialized */
|
||||||
|
int num_registers; /* number of registers used */
|
||||||
|
unsigned char anchor; /* anchor: 0=none 1=begline 2=begbuf */
|
||||||
|
} *regexp_t;
|
||||||
|
|
||||||
|
typedef struct re_registers
|
||||||
|
{
|
||||||
|
int start[RE_NREGS]; /* start offset of region */
|
||||||
|
int end[RE_NREGS]; /* end offset of region */
|
||||||
|
} *regexp_registers_t;
|
||||||
|
|
||||||
|
/* bit definitions for syntax */
|
||||||
|
#define RE_NO_BK_PARENS 1 /* no quoting for parentheses */
|
||||||
|
#define RE_NO_BK_VBAR 2 /* no quoting for vertical bar */
|
||||||
|
#define RE_BK_PLUS_QM 4 /* quoting needed for + and ? */
|
||||||
|
#define RE_TIGHT_VBAR 8 /* | binds tighter than ^ and $ */
|
||||||
|
#define RE_NEWLINE_OR 16 /* treat newline as or */
|
||||||
|
#define RE_CONTEXT_INDEP_OPS 32 /* ^$?*+ are special in all contexts */
|
||||||
|
#define RE_ANSI_HEX 64 /* ansi sequences (\n etc) and \xhh */
|
||||||
|
#define RE_NO_GNU_EXTENSIONS 128 /* no gnu extensions */
|
||||||
|
|
||||||
|
#define TP_RE_NOERR 0
|
||||||
|
#define TP_RE_UNKNOWN_OPCODE (-1)
|
||||||
|
#define TP_RE_JUMP_OUT_BOUNDS 1
|
||||||
|
#define TP_RE_QUOTE_ERR 2
|
||||||
|
|
||||||
|
/* definitions for some common regexp styles */
|
||||||
|
#define RE_SYNTAX_AWK (RE_NO_BK_PARENS|RE_NO_BK_VBAR|RE_CONTEXT_INDEP_OPS)
|
||||||
|
#define RE_SYNTAX_EGREP (RE_SYNTAX_AWK|RE_NEWLINE_OR)
|
||||||
|
#define RE_SYNTAX_GREP (RE_BK_PLUS_QM|RE_NEWLINE_OR)
|
||||||
|
#define RE_SYNTAX_EMACS 0
|
||||||
|
|
||||||
|
#define Sword 1
|
||||||
|
#define Swhitespace 2
|
||||||
|
#define Sdigit 4
|
||||||
|
#define Soctaldigit 8
|
||||||
|
#define Shexdigit 16
|
||||||
|
|
||||||
|
/* Rename all exported symbols to avoid conflicts with similarly named
|
||||||
|
symbols in some systems' standard C libraries... */
|
||||||
|
|
||||||
|
#define re_syntax _Py_re_syntax
|
||||||
|
#define re_syntax_table _Py_re_syntax_table
|
||||||
|
#define re_compile_initialize _Py_re_compile_initialize
|
||||||
|
#define re_set_syntax _Py_re_set_syntax
|
||||||
|
#define re_compile_pattern _Py_re_compile_pattern
|
||||||
|
#define re_match _Py_re_match
|
||||||
|
#define re_search _Py_re_search
|
||||||
|
#define re_compile_fastmap _Py_re_compile_fastmap
|
||||||
|
#define re_comp _Py_re_comp
|
||||||
|
#define re_exec _Py_re_exec
|
||||||
|
|
||||||
|
#ifdef HAVE_PROTOTYPES
|
||||||
|
|
||||||
|
extern int re_syntax;
|
||||||
|
/* This is the actual syntax mask. It was added so that Python could do
|
||||||
|
* syntax-dependent munging of patterns before compilation. */
|
||||||
|
|
||||||
|
extern unsigned char re_syntax_table[256];
|
||||||
|
|
||||||
|
void re_compile_initialize(void);
|
||||||
|
|
||||||
|
int re_set_syntax(int syntax);
|
||||||
|
/* This sets the syntax to use and returns the previous syntax. The
|
||||||
|
* syntax is specified by a bit mask of the above defined bits. */
|
||||||
|
|
||||||
|
char *re_compile_pattern(unsigned char *regex, int regex_size, regexp_t compiled);
|
||||||
|
/* This compiles the regexp (given in regex and length in regex_size).
|
||||||
|
* This returns NULL if the regexp compiled successfully, and an error
|
||||||
|
* message if an error was encountered. The buffer field must be
|
||||||
|
* initialized to a memory area allocated by malloc (or to NULL) before
|
||||||
|
* use, and the allocated field must be set to its length (or 0 if
|
||||||
|
* buffer is NULL). Also, the translate field must be set to point to a
|
||||||
|
* valid translation table, or NULL if it is not used. */
|
||||||
|
|
||||||
|
int re_match(regexp_t compiled, unsigned char *string, int size, int pos,
|
||||||
|
regexp_registers_t old_regs);
|
||||||
|
/* This tries to match the regexp against the string. This returns the
|
||||||
|
* length of the matched portion, or -1 if the pattern could not be
|
||||||
|
* matched and -2 if an error (such as failure stack overflow) is
|
||||||
|
* encountered. */
|
||||||
|
|
||||||
|
int re_search(regexp_t compiled, unsigned char *string, int size, int startpos,
|
||||||
|
int range, regexp_registers_t regs);
|
||||||
|
/* This searches for a substring matching the regexp. This returns the
|
||||||
|
* first index at which a match is found. range specifies at how many
|
||||||
|
* positions to try matching; positive values indicate searching
|
||||||
|
* forwards, and negative values indicate searching backwards. mstop
|
||||||
|
* specifies the offset beyond which a match must not go. This returns
|
||||||
|
* -1 if no match is found, and -2 if an error (such as failure stack
|
||||||
|
* overflow) is encountered. */
|
||||||
|
|
||||||
|
void re_compile_fastmap(regexp_t compiled);
|
||||||
|
/* This computes the fastmap for the regexp. For this to have any effect,
|
||||||
|
* the calling program must have initialized the fastmap field to point
|
||||||
|
* to an array of 256 characters. */
|
||||||
|
|
||||||
|
#else /* HAVE_PROTOTYPES */
|
||||||
|
|
||||||
|
extern int re_syntax;
|
||||||
|
extern unsigned char re_syntax_table[256];
|
||||||
|
void re_compile_initialize();
|
||||||
|
int re_set_syntax();
|
||||||
|
char *re_compile_pattern();
|
||||||
|
int re_match();
|
||||||
|
int re_search();
|
||||||
|
void re_compile_fastmap();
|
||||||
|
|
||||||
|
#endif /* HAVE_PROTOTYPES */
|
||||||
|
|
||||||
|
#endif /* REGEXPR_H */
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#ifdef __cplusplus
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#endif /* !Py_REGEXPR_H */
|
648
programs/develop/tinypy/modules/re/tests.py
Normal file
648
programs/develop/tinypy/modules/re/tests.py
Normal file
@@ -0,0 +1,648 @@
|
|||||||
|
"""
|
||||||
|
test case for re module
|
||||||
|
"""
|
||||||
|
|
||||||
|
import re
|
||||||
|
import testsuite
|
||||||
|
SUCCEED, FAIL, SYNTAX_ERROR = range(3)
|
||||||
|
|
||||||
|
def RAISE():
|
||||||
|
raise("testing failed")
|
||||||
|
|
||||||
|
def main():
|
||||||
|
#print("begin re tests")
|
||||||
|
|
||||||
|
assert(re.__name__ != None)
|
||||||
|
assert(re.__doc__ != None)
|
||||||
|
assert(re.__file__ != None)
|
||||||
|
|
||||||
|
test_re_obj_search()
|
||||||
|
test_re_obj_match()
|
||||||
|
test_re_mod_search()
|
||||||
|
test_re_mod_match()
|
||||||
|
test_re_obj_split()
|
||||||
|
test_re_mod_split()
|
||||||
|
test_re_obj_findall()
|
||||||
|
test_re_mod_findall()
|
||||||
|
test_mat_obj_groups()
|
||||||
|
test_mat_obj_start()
|
||||||
|
test_mat_obj_end()
|
||||||
|
test_mat_obj_span()
|
||||||
|
|
||||||
|
print("#OK: re tests passed")
|
||||||
|
|
||||||
|
def test_re_obj_search(verbose = None):
|
||||||
|
"""
|
||||||
|
some tests borrowed from cpython
|
||||||
|
testing re.compile(), reobj.search(), and matobj.group()
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.search_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern=s=outcome=repl=expected=None
|
||||||
|
if len(t)==5:
|
||||||
|
pattern, s, outcome, repl, expected = t
|
||||||
|
elif len(t)==3:
|
||||||
|
pattern, s, outcome = t
|
||||||
|
else:
|
||||||
|
raise ('Test tuples should have 3 or 5 fields',t)
|
||||||
|
|
||||||
|
try:
|
||||||
|
obj=re.compile(pattern)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR: continue # Expected a syntax error
|
||||||
|
else:
|
||||||
|
# Regex syntax errors aren't yet reported, so for
|
||||||
|
# the official test suite they'll be quietly ignored.
|
||||||
|
pass
|
||||||
|
try:
|
||||||
|
matobj=obj.search(s)
|
||||||
|
except:
|
||||||
|
print('=== Unexpected exception:', obj, matobj, pattern, s)
|
||||||
|
RAISE()
|
||||||
|
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
pass
|
||||||
|
elif outcome==FAIL:
|
||||||
|
if matobj==None: pass # No match, as expected
|
||||||
|
else: print('=== Succeeded incorrectly', obj, matobj, pattern, s)
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if matobj!=None:
|
||||||
|
# Matched, as expected, so now we compute the
|
||||||
|
# result string and compare it to our expected result.
|
||||||
|
found=matobj.group(0)
|
||||||
|
repl = repl.replace("found", str(found))
|
||||||
|
for i in range(1,11):
|
||||||
|
if "g"+str(i) in repl:
|
||||||
|
gi = str(matobj.group(i))
|
||||||
|
repl = repl.replace("g"+str(i), gi)
|
||||||
|
if len(t) == 5:
|
||||||
|
repl = repl.replace('+', '')
|
||||||
|
repl = repl.replace('\"', '')
|
||||||
|
if repl!=expected:
|
||||||
|
print( '=== grouping error', t,
|
||||||
|
str(repl)+' should be '+str(expected))
|
||||||
|
RAISE()
|
||||||
|
else:
|
||||||
|
print ('=== Failed incorrectly', t)
|
||||||
|
|
||||||
|
def test_re_obj_match(verbose = None):
|
||||||
|
"""
|
||||||
|
some tests borrowed from cpython
|
||||||
|
testing re.compile(), reobj.match() and matobj.group()
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.match_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern=s=outcome=repl=expected=None
|
||||||
|
if len(t)==5:
|
||||||
|
pattern, s, outcome, repl, expected = t
|
||||||
|
elif len(t)==3:
|
||||||
|
pattern, s, outcome = t
|
||||||
|
else:
|
||||||
|
raise ('Test tuples should have 3 or 5 fields',t)
|
||||||
|
|
||||||
|
try:
|
||||||
|
obj=re.compile(pattern)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR: continue # Expected a syntax error
|
||||||
|
else:
|
||||||
|
# Regex syntax errors aren't yet reported, so for
|
||||||
|
# the official test suite they'll be quietly ignored.
|
||||||
|
pass
|
||||||
|
try:
|
||||||
|
matobj=obj.match(s)
|
||||||
|
except:
|
||||||
|
print('=== Unexpected exception:', obj, matobj, pattern, s)
|
||||||
|
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
pass
|
||||||
|
elif outcome==FAIL:
|
||||||
|
if matobj==None: pass # No match, as expected
|
||||||
|
else: print('=== Succeeded incorrectly', obj, matobj, pattern, s)
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if matobj!=None:
|
||||||
|
# Matched, as expected, so now we compute the
|
||||||
|
# result string and compare it to our expected result.
|
||||||
|
found=matobj.group(0)
|
||||||
|
repl = repl.replace("found", str(found))
|
||||||
|
for i in range(1,11):
|
||||||
|
if "g"+str(i) in repl:
|
||||||
|
gi = str(matobj.group(i))
|
||||||
|
repl = repl.replace("g"+str(i), gi)
|
||||||
|
if len(t) == 5:
|
||||||
|
repl = repl.replace('+', '')
|
||||||
|
repl = repl.replace('\"', '')
|
||||||
|
if repl!=expected:
|
||||||
|
print( '=== grouping error', t,
|
||||||
|
str(repl)+' should be '+str(expected))
|
||||||
|
RAISE()
|
||||||
|
else:
|
||||||
|
print ('=== Failed incorrectly', obj, matobj, pattern, s)
|
||||||
|
|
||||||
|
def test_re_mod_search(verbose = None):
|
||||||
|
"""
|
||||||
|
some tests borrowed from cpython
|
||||||
|
testing re.search(), and matobj.group()
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.search_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern=s=outcome=repl=expected=None
|
||||||
|
if len(t)==5:
|
||||||
|
pattern, s, outcome, repl, expected = t
|
||||||
|
elif len(t)==3:
|
||||||
|
pattern, s, outcome = t
|
||||||
|
else:
|
||||||
|
raise ('Test tuples should have 3 or 5 fields',t)
|
||||||
|
|
||||||
|
try:
|
||||||
|
matobj=re.search(pattern, s)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception:', matobj, pattern, s)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
if matobj==None: pass # No match, as expected
|
||||||
|
else: print('=== Succeeded incorrectly', obj, matobj, pattern, s)
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if matobj!=None:
|
||||||
|
# Matched, as expected, so now we compute the
|
||||||
|
# result string and compare it to our expected result.
|
||||||
|
found=matobj.group(0)
|
||||||
|
repl = repl.replace("found", str(found))
|
||||||
|
for i in range(1,11):
|
||||||
|
if "g"+str(i) in repl:
|
||||||
|
gi = str(matobj.group(i))
|
||||||
|
repl = repl.replace("g"+str(i), gi)
|
||||||
|
if len(t) == 5:
|
||||||
|
repl = repl.replace('+', '')
|
||||||
|
repl = repl.replace('\"', '')
|
||||||
|
if repl!=expected:
|
||||||
|
print( '=== grouping error', t,
|
||||||
|
str(repl)+' should be '+str(expected))
|
||||||
|
RAISE()
|
||||||
|
else:
|
||||||
|
print ('=== Failed incorrectly', t)
|
||||||
|
|
||||||
|
def test_re_mod_match(verbose = None):
|
||||||
|
"""
|
||||||
|
some tests borrowed from cpython
|
||||||
|
testing re.match(), and matobj.group()
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.match_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern=s=outcome=repl=expected=None
|
||||||
|
if len(t)==5:
|
||||||
|
pattern, s, outcome, repl, expected = t
|
||||||
|
elif len(t)==3:
|
||||||
|
pattern, s, outcome = t
|
||||||
|
else:
|
||||||
|
raise ('Test tuples should have 3 or 5 fields',t)
|
||||||
|
|
||||||
|
try:
|
||||||
|
matobj=re.match(pattern, s)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception:', matobj, pattern, s)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
if matobj==None: pass # No match, as expected
|
||||||
|
else: print('=== Succeeded incorrectly', matobj, pattern, s)
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if matobj!=None:
|
||||||
|
# Matched, as expected, so now we compute the
|
||||||
|
# result string and compare it to our expected result.
|
||||||
|
found=matobj.group(0)
|
||||||
|
repl = repl.replace("found", str(found))
|
||||||
|
for i in range(1,11):
|
||||||
|
if "g"+str(i) in repl:
|
||||||
|
gi = str(matobj.group(i))
|
||||||
|
repl = repl.replace("g"+str(i), gi)
|
||||||
|
if len(t) == 5:
|
||||||
|
repl = repl.replace('+', '')
|
||||||
|
repl = repl.replace('\"', '')
|
||||||
|
if repl!=expected:
|
||||||
|
print( '=== grouping error', t,
|
||||||
|
str(repl)+' should be '+str(expected))
|
||||||
|
RAISE()
|
||||||
|
else:
|
||||||
|
print ('=== Failed incorrectly', t)
|
||||||
|
|
||||||
|
def test_re_obj_split(verbose = None):
|
||||||
|
"""
|
||||||
|
test re.compile(), and reobj.split()
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.split_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern, s, outcome, maxsplit, fields = t
|
||||||
|
try:
|
||||||
|
reobj = re.compile(pattern)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception:', pattern, s,
|
||||||
|
outcome, maxsplit, fields)
|
||||||
|
try:
|
||||||
|
fldlst=reobj.split(s, maxsplit)
|
||||||
|
except:
|
||||||
|
if outcome == SYNTAX_ERROR:
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception:', pattern, s,
|
||||||
|
outcome, maxsplit, fields)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
pass # No match, as expected
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if fldlst:
|
||||||
|
# Matched, as expected, so now we compute the
|
||||||
|
# result string and compare it to our expected result.
|
||||||
|
if verbose:
|
||||||
|
fldstr = fieldstr = ""
|
||||||
|
for item in fldlst:
|
||||||
|
fldstr = fldstr + str(item) + " | "
|
||||||
|
for item in fields:
|
||||||
|
fieldstr = fieldstr + str(item) + " | "
|
||||||
|
print(fldstr, "~~~", fieldstr)
|
||||||
|
if len(fields) != len(fldlst):
|
||||||
|
print('=== Not coherent 1')
|
||||||
|
RAISE()
|
||||||
|
|
||||||
|
for i in range(len(fields)):
|
||||||
|
if fields[i] != fldlst[i]:
|
||||||
|
if verbose:
|
||||||
|
print('=== Not coherent 2', pattern, s,
|
||||||
|
outcome, maxsplit, fields, i,
|
||||||
|
fields[i],'(',len(fields[i]),')', ' | ',
|
||||||
|
fldlst[i],'(',len(fldlst[i]),')')
|
||||||
|
else:
|
||||||
|
print('=== Not coherent 2')
|
||||||
|
RAISE()
|
||||||
|
else:
|
||||||
|
print ('=== Failed incorrectly', pattern, s,
|
||||||
|
outcome, maxsplit, fields)
|
||||||
|
|
||||||
|
def test_re_mod_split(verbose = None):
|
||||||
|
"""
|
||||||
|
test re.split()
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.split_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern, s, outcome, maxsplit, fields = t
|
||||||
|
try:
|
||||||
|
fldlst=re.split(pattern, s, maxsplit)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception:', pattern, s,
|
||||||
|
outcome, maxsplit, fields)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
pass # No match, as expected
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if fldlst:
|
||||||
|
# Matched, as expected, so now we compute the
|
||||||
|
# result string and compare it to our expected result.
|
||||||
|
if verbose:
|
||||||
|
fldstr = fieldstr = ""
|
||||||
|
for item in fldlst:
|
||||||
|
fldstr = fldstr + str(item) + " | "
|
||||||
|
for item in fields:
|
||||||
|
fieldstr = fieldstr + str(item) + " | "
|
||||||
|
print(fldstr, "~~~", fieldstr)
|
||||||
|
|
||||||
|
if len(fields) != len(fldlst):
|
||||||
|
print('=== Not coherent 1')
|
||||||
|
RAISE()
|
||||||
|
|
||||||
|
for i in range(len(fields)):
|
||||||
|
if fields[i] != fldlst[i]:
|
||||||
|
if verbose:
|
||||||
|
print('=== Not coherent 2', pattern, s,
|
||||||
|
outcome, maxsplit, fields, i,
|
||||||
|
fields[i],'(',len(fields[i]),')', ' | ',
|
||||||
|
fldlst[i],'(',len(fldlst[i]),')')
|
||||||
|
else:
|
||||||
|
print('=== Not coherent 2')
|
||||||
|
RAISE()
|
||||||
|
else:
|
||||||
|
print ('=== Failed incorrectly', pattern, s,
|
||||||
|
outcome, maxsplit, fields)
|
||||||
|
|
||||||
|
def test_re_obj_findall(verbose = None):
|
||||||
|
"""
|
||||||
|
test re.compile(), and reobj.findall()
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.findall_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern, s, outcome, pos, fields = t
|
||||||
|
try:
|
||||||
|
reobj = re.compile(pattern)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception:', pattern, s,
|
||||||
|
outcome, pos, fields)
|
||||||
|
try:
|
||||||
|
fldlst=reobj.findall(s, pos)
|
||||||
|
except:
|
||||||
|
if outcome == SYNTAX_ERROR:
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception:', pattern, s,
|
||||||
|
outcome, pos, fields)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
pass # No match, as expected
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if fldlst:
|
||||||
|
# Matched, as expected, so now we compute the
|
||||||
|
# result string and compare it to our expected result.
|
||||||
|
if verbose:
|
||||||
|
fldstr = fieldstr = ""
|
||||||
|
for item in fldlst:
|
||||||
|
fldstr = fldstr + str(item) + " | "
|
||||||
|
for item in fields:
|
||||||
|
fieldstr = fieldstr + str(item) + " | "
|
||||||
|
print(fldstr, "~~~", fieldstr)
|
||||||
|
|
||||||
|
if len(fields) != len(fldlst):
|
||||||
|
print('=== Not coherent 1')
|
||||||
|
RAISE()
|
||||||
|
|
||||||
|
for i in range(len(fields)):
|
||||||
|
if fields[i] != fldlst[i]:
|
||||||
|
if verbose:
|
||||||
|
print('=== Not coherent 2', pattern, s,
|
||||||
|
outcome, maxsplit, fields, i,
|
||||||
|
fields[i],'(',len(fields[i]),')', ' | ',
|
||||||
|
fldlst[i],'(',len(fldlst[i]),')')
|
||||||
|
else:
|
||||||
|
print('=== Not coherent 2')
|
||||||
|
RAISE()
|
||||||
|
else:
|
||||||
|
print ('=== Failed incorrectly', pattern, s,
|
||||||
|
outcome, pos, fields)
|
||||||
|
|
||||||
|
def test_re_mod_findall(verbose = None):
|
||||||
|
"""
|
||||||
|
test re.findall()
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.mod_findall_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern, s, outcome, pos, fields = t # pos is not used
|
||||||
|
try:
|
||||||
|
fldlst=re.findall(pattern, s)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception:', pattern, s,
|
||||||
|
outcome, pos, fields)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
pass # No match, as expected
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if fldlst:
|
||||||
|
# Matched, as expected, so now we compute the
|
||||||
|
# result string and compare it to our expected result.
|
||||||
|
if verbose:
|
||||||
|
fldstr = fieldstr = ""
|
||||||
|
for item in fldlst:
|
||||||
|
fldstr = fldstr + str(item) + " | "
|
||||||
|
for item in fields:
|
||||||
|
fieldstr = fieldstr + str(item) + " | "
|
||||||
|
print(fldstr, "~~~", fieldstr)
|
||||||
|
|
||||||
|
if len(fields) != len(fldlst):
|
||||||
|
print('=== Not coherent 1')
|
||||||
|
RAISE()
|
||||||
|
|
||||||
|
for i in range(len(fields)):
|
||||||
|
if fields[i] != fldlst[i]:
|
||||||
|
if verbose:
|
||||||
|
print('=== Not coherent 2', pattern, s,
|
||||||
|
outcome, maxsplit, fields, i,
|
||||||
|
fields[i],'(',len(fields[i]),')', ' | ',
|
||||||
|
fldlst[i],'(',len(fldlst[i]),')')
|
||||||
|
else:
|
||||||
|
print('=== Not coherent 2')
|
||||||
|
RAISE()
|
||||||
|
else:
|
||||||
|
print ('=== Failed incorrectly', pattern, s,
|
||||||
|
outcome, pos, fields)
|
||||||
|
|
||||||
|
def test_mat_obj_groups(verbose = None):
|
||||||
|
"""
|
||||||
|
test re.search(), and matobj.groups()
|
||||||
|
'verbose' is for debugging, when 'verbose' is true, print extra info
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.matobj_groups_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern, s, outcome, fields, grpidx, start, end = t
|
||||||
|
try:
|
||||||
|
matobj=re.search(pattern, s)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception 1:', pattern, s,
|
||||||
|
outcome,fields)
|
||||||
|
|
||||||
|
try:
|
||||||
|
if outcome==SUCCEED: assert(matobj != None)
|
||||||
|
fldlst = matobj.groups()
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception 2:', pattern, s,
|
||||||
|
outcome,fields)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
pass # No match, as expected
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if fldlst and fields:
|
||||||
|
# Matched, as expected, so now we compute the
|
||||||
|
# result string and compare it to our expected result.
|
||||||
|
if verbose:
|
||||||
|
fldstr = fieldstr = ""
|
||||||
|
for item in fldlst:
|
||||||
|
fldstr = fldstr + str(item) + " | "
|
||||||
|
for item in fields:
|
||||||
|
fieldstr = fieldstr + str(item) + " | "
|
||||||
|
print(fldstr, "~~~", fieldstr)
|
||||||
|
|
||||||
|
if len(fields) != len(fldlst):
|
||||||
|
print('=== Not coherent 2')
|
||||||
|
RAISE()
|
||||||
|
|
||||||
|
for i in range(len(fields)):
|
||||||
|
if fields[i] != fldlst[i]:
|
||||||
|
if verbose:
|
||||||
|
print('=== Not coherent', pattern, s,
|
||||||
|
outcome,fields, i,
|
||||||
|
fields[i],'(',len(fields[i]),')', ' | ',
|
||||||
|
fldlst[i],'(',len(fldlst[i]),')')
|
||||||
|
else:
|
||||||
|
print('=== Not coherent')
|
||||||
|
RAISE()
|
||||||
|
elif not len(fldlst) and not len(fields):
|
||||||
|
# output is empty, as expected
|
||||||
|
if verbose:
|
||||||
|
print("output is empty, as expected")
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
if verbose:
|
||||||
|
for item in fldlst:
|
||||||
|
print(item,)
|
||||||
|
print()
|
||||||
|
for item in fields:
|
||||||
|
print(item,)
|
||||||
|
print()
|
||||||
|
print ('=== Failed incorrectly', pattern, s,
|
||||||
|
outcome,fields,fldlst)
|
||||||
|
|
||||||
|
def test_mat_obj_start(verbose = None):
|
||||||
|
"""
|
||||||
|
test re.search(), and matobj.start()
|
||||||
|
'verbose' is for debugging, when 'verbose' is true, print extra info
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.matobj_groups_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern, s, outcome, fields, grpidx, start, end = t
|
||||||
|
try:
|
||||||
|
matobj=re.search(pattern, s)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception 1:', pattern, s,
|
||||||
|
outcome,fields)
|
||||||
|
|
||||||
|
try:
|
||||||
|
if outcome==SUCCEED: assert(matobj != None)
|
||||||
|
fldlst = matobj.groups()
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception 2:', pattern, s,
|
||||||
|
outcome,fields)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
pass # No match, as expected
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if grpidx > 0:
|
||||||
|
if matobj.start(grpidx) == start:
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
if verbose:
|
||||||
|
print ('=== Failed incorrectly', pattern, s,
|
||||||
|
outcome,fields,fldlst)
|
||||||
|
raise("testing failed")
|
||||||
|
|
||||||
|
|
||||||
|
def test_mat_obj_end(verbose = None):
|
||||||
|
"""
|
||||||
|
test re.search(), and matobj.end()
|
||||||
|
'verbose' is for debugging, when 'verbose' is true, print extra info
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.matobj_groups_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern, s, outcome, fields, grpidx, start, end = t
|
||||||
|
try:
|
||||||
|
matobj=re.search(pattern, s)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception 1:', pattern, s,
|
||||||
|
outcome,fields)
|
||||||
|
|
||||||
|
try:
|
||||||
|
if outcome==SUCCEED: assert(matobj != None)
|
||||||
|
fldlst = matobj.groups()
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception 2:', pattern, s,
|
||||||
|
outcome,fields)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
pass # No match, as expected
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if grpidx > 0:
|
||||||
|
if matobj.end(grpidx) == end:
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
if verbose:
|
||||||
|
print ('=== Failed incorrectly', pattern, s,
|
||||||
|
outcome,fields,fldlst, matobj.end(grpidx), end)
|
||||||
|
raise("testing failed")
|
||||||
|
|
||||||
|
def test_mat_obj_span(verbose = None):
|
||||||
|
"""
|
||||||
|
test re.search(), and matobj.span()
|
||||||
|
'verbose' is for debugging, when 'verbose' is true, print extra info
|
||||||
|
"""
|
||||||
|
regex_tests = testsuite.matobj_groups_regex_tests
|
||||||
|
for t in regex_tests:
|
||||||
|
pattern, s, outcome, fields, grpidx, start, end = t
|
||||||
|
try:
|
||||||
|
matobj=re.search(pattern, s)
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception 1:', pattern, s,
|
||||||
|
outcome,fields)
|
||||||
|
|
||||||
|
try:
|
||||||
|
if outcome==SUCCEED: assert(matobj != None)
|
||||||
|
fldlst = matobj.groups()
|
||||||
|
except:
|
||||||
|
if outcome==SYNTAX_ERROR:
|
||||||
|
# This should have been a syntax error; forget it.
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
print('=== Unexpected exception 2:', pattern, s,
|
||||||
|
outcome,fields)
|
||||||
|
|
||||||
|
if outcome==FAIL:
|
||||||
|
pass # No match, as expected
|
||||||
|
elif outcome==SUCCEED:
|
||||||
|
if (grpidx > 0):
|
||||||
|
spstart, spend = matobj.span(grpidx)
|
||||||
|
if spstart == start and spend == end:
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
if verbose:
|
||||||
|
print ('=== Failed incorrectly', pattern, s,
|
||||||
|
outcome,fields,fldlst)
|
||||||
|
raise("testing failed")
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
|
|
367
programs/develop/tinypy/modules/re/testsuite.py
Normal file
367
programs/develop/tinypy/modules/re/testsuite.py
Normal file
@@ -0,0 +1,367 @@
|
|||||||
|
# Test suite (for verifying correctness)
|
||||||
|
#
|
||||||
|
# The test suite is a list of 5- or 3-tuples. The 5 parts of a
|
||||||
|
# complete tuple are:
|
||||||
|
# element 0: a string containing the pattern
|
||||||
|
# 1: the string to match against the pattern
|
||||||
|
# 2: the expected result (0 - SUCCEED, 1 - FAIL, 2 - SYNTAX_ERROR)
|
||||||
|
# 3: a string that will be eval()'ed to produce a test string.
|
||||||
|
# This is an arbitrary Python expression; the available
|
||||||
|
# variables are "found" (the whole match), and "g1", "g2", ...
|
||||||
|
# up to "g10" contain the contents of each group, or the
|
||||||
|
# string 'None' if the group wasn't given a value.
|
||||||
|
# 4: The expected result of evaluating the expression.
|
||||||
|
# If the two don't match, an error is reported.
|
||||||
|
#
|
||||||
|
# If the regex isn't expected to work, the latter two elements can be omitted.
|
||||||
|
|
||||||
|
# test suite for search
|
||||||
|
search_regex_tests=[
|
||||||
|
['abc', 'abc', 0, 'found', 'abc'],
|
||||||
|
['abc', 'xbc', 1],
|
||||||
|
['abc', 'axc', 1],
|
||||||
|
['abc', 'abx', 1],
|
||||||
|
['abc', 'xabcy', 0, 'found', 'abc'],
|
||||||
|
['abc', 'ababc', 0, 'found', 'abc'],
|
||||||
|
['ab*c', 'abc', 0, 'found', 'abc'],
|
||||||
|
['ab*bc', 'abc', 0, 'found', 'abc'],
|
||||||
|
['ab*bc', 'abbc', 0, 'found', 'abbc'],
|
||||||
|
['ab*bc', 'abbbbc', 0, 'found', 'abbbbc'],
|
||||||
|
['ab+bc', 'abbc', 0, 'found', 'abbc'],
|
||||||
|
['ab+bc', 'abc', 1],
|
||||||
|
['ab+bc', 'abq', 1],
|
||||||
|
['ab+bc', 'abbbbc', 0, 'found', 'abbbbc'],
|
||||||
|
['ab?bc', 'abbc', 0, 'found', 'abbc'],
|
||||||
|
['ab?bc', 'abc', 0, 'found', 'abc'],
|
||||||
|
['ab?bc', 'abbbbc', 1],
|
||||||
|
['ab?c', 'abc', 0, 'found', 'abc'],
|
||||||
|
['^abc$', 'abc', 0, 'found', 'abc'],
|
||||||
|
['^abc$', 'abcc', 1],
|
||||||
|
['^abc', 'abcc', 0, 'found', 'abc'],
|
||||||
|
['^abc$', 'aabc', 1],
|
||||||
|
['abc$', 'aabc', 0, 'found', 'abc'],
|
||||||
|
['^', 'abc', 0, 'found+"-"', '-'],
|
||||||
|
['$', 'abc', 0, 'found+"-"', '-'],
|
||||||
|
['a.c', 'abc', 0, 'found', 'abc'],
|
||||||
|
['a.c', 'axc', 0, 'found', 'axc'],
|
||||||
|
['a.*c', 'axyzc', 0, 'found', 'axyzc'],
|
||||||
|
['a.*c', 'axyzd', 1],
|
||||||
|
['a[bc]d', 'abc', 1],
|
||||||
|
['a[bc]d', 'abd', 0, 'found', 'abd'],
|
||||||
|
['a[b-d]e', 'abd', 1],
|
||||||
|
['a[b-d]e', 'ace', 0, 'found', 'ace'],
|
||||||
|
['a[b-d]', 'aac', 0, 'found', 'ac'],
|
||||||
|
['a[-b]', 'a-', 0, 'found', 'a-'],
|
||||||
|
['a[b-]', 'a-', 0, 'found', 'a-'],
|
||||||
|
['a[]b', '-', 2],
|
||||||
|
['a[', '-', 2],
|
||||||
|
['a\\', '-', 2],
|
||||||
|
['abc\\)', '-', 2],
|
||||||
|
['\\(abc', '-', 2],
|
||||||
|
['a]', 'a]', 0, 'found', 'a]'],
|
||||||
|
['a[]]b', 'a]b', 0, 'found', 'a]b'],
|
||||||
|
['a[^bc]d', 'aed', 0, 'found', 'aed'],
|
||||||
|
['a[^bc]d', 'abd', 1],
|
||||||
|
['a[^-b]c', 'adc', 0, 'found', 'adc'],
|
||||||
|
['a[^-b]c', 'a-c', 1],
|
||||||
|
['a[^]b]c', 'a]c', 1],
|
||||||
|
['a[^]b]c', 'adc', 0, 'found', 'adc'],
|
||||||
|
['\\ba\\b', 'a-', 0, '"-"', '-'],
|
||||||
|
['\\ba\\b', '-a', 0, '"-"', '-'],
|
||||||
|
['\\ba\\b', '-a-', 0, '"-"', '-'],
|
||||||
|
['\\by\\b', 'xy', 1],
|
||||||
|
['\\by\\b', 'yz', 1],
|
||||||
|
['\\by\\b', 'xyz', 1],
|
||||||
|
['ab\\|cd', 'abc', 0, 'found', 'ab'],
|
||||||
|
['ab\\|cd', 'abcd', 0, 'found', 'ab'],
|
||||||
|
['\\(\\)ef', 'def', 0, 'found+"-"+g1', 'ef-'],
|
||||||
|
['$b', 'b', 1],
|
||||||
|
['a(b', 'a(b', 0, 'found+"-"+g1', 'a(b-None'],
|
||||||
|
['a(*b', 'ab', 0, 'found', 'ab'],
|
||||||
|
['a(*b', 'a((b', 0, 'found', 'a((b'],
|
||||||
|
['a\\\\b', 'a\\b', 0, 'found', 'a\\b'],
|
||||||
|
['\\(\\(a\\)\\)', 'abc', 0, 'found+"-"+g1+"-"+g2', 'a-a-a'],
|
||||||
|
['\\(a\\)b\\(c\\)', 'abc', 0, 'found+"-"+g1+"-"+g2', 'abc-a-c'],
|
||||||
|
['a+b+c', 'aabbabc', 0, 'found', 'abc'],
|
||||||
|
['\\(a+\\|b\\)*', 'ab', 0, 'found+"-"+g1', 'ab-b'],
|
||||||
|
['\\(a+\\|b\\)+', 'ab', 0, 'found+"-"+g1', 'ab-b'],
|
||||||
|
['\\(a+\\|b\\)?', 'ab', 0, 'found+"-"+g1', 'a-a'],
|
||||||
|
['\\)\\(', '-', 2],
|
||||||
|
['[^ab]*', 'cde', 0, 'found', 'cde'],
|
||||||
|
['abc', '', 1],
|
||||||
|
['a*', '', 0, 'found', ''],
|
||||||
|
['a\\|b\\|c\\|d\\|e', 'e', 0, 'found', 'e'],
|
||||||
|
['\\(a\\|b\\|c\\|d\\|e\\)f', 'ef', 0, 'found+"-"+g1', 'ef-e'],
|
||||||
|
['abcd*efg', 'abcdefg', 0, 'found', 'abcdefg'],
|
||||||
|
['ab*', 'xabyabbbz', 0, 'found', 'ab'],
|
||||||
|
['ab*', 'xayabbbz', 0, 'found', 'a'],
|
||||||
|
['\\(ab\\|cd\\)e', 'abcde', 0, 'found+"-"+g1', 'cde-cd'],
|
||||||
|
['[abhgefdc]ij', 'hij', 0, 'found', 'hij'],
|
||||||
|
['^\\(ab\\|cd\\)e', 'abcde', 1, 'xg1y', 'xy'],
|
||||||
|
['\\(abc\\|\\)ef', 'abcdef', 0, 'found+"-"+g1', 'ef-'],
|
||||||
|
['\\(a\\|b\\)c*d', 'abcd', 0, 'found+"-"+g1', 'bcd-b'],
|
||||||
|
['\\(ab\\|ab*\\)bc', 'abc', 0, 'found+"-"+g1', 'abc-a'],
|
||||||
|
['a\\([bc]*\\)c*', 'abc', 0, 'found+"-"+g1', 'abc-bc'],
|
||||||
|
['a\\([bc]*\\)\\(c*d\\)', 'abcd', 0, 'found+"-"+g1+"-"+g2', 'abcd-bc-d'],
|
||||||
|
['a\\([bc]+\\)\\(c*d\\)', 'abcd', 0, 'found+"-"+g1+"-"+g2', 'abcd-bc-d'],
|
||||||
|
['a\\([bc]*\\)\\(c+d\\)', 'abcd', 0, 'found+"-"+g1+"-"+g2', 'abcd-b-cd'],
|
||||||
|
['a[bcd]*dcdcde', 'adcdcde', 0, 'found', 'adcdcde'],
|
||||||
|
['a[bcd]+dcdcde', 'adcdcde', 1],
|
||||||
|
['\\(ab\\|a\\)b*c', 'abc', 0, 'found+"-"+g1', 'abc-ab'],
|
||||||
|
['\\(\\(a\\)\\(b\\)c\\)\\(d\\)', 'abcd', 0, 'g1+"-"+g2+"-"+g3+"-"+g4', 'abc-a-b-d'],
|
||||||
|
['[a-zA-Z_][a-zA-Z0-9_]*', 'alpha', 0, 'found', 'alpha'],
|
||||||
|
['^a\\(bc+\\|b[eh]\\)g\\|.h$', 'abh', 0, 'found+"-"+g1', 'bh-None'],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'effgz', 0, 'found+"-"+g1+"-"+g2', 'effgz-effgz-None'],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'ij', 0, 'found+"-"+g1+"-"+g2', 'ij-ij-j'],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'effg', 1],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'bcdd', 1],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'reffgz', 0, 'found+"-"+g1+"-"+g2', 'effgz-effgz-None'],
|
||||||
|
['\\(\\(\\(\\(\\(\\(\\(\\(\\(a\\)\\)\\)\\)\\)\\)\\)\\)\\)', 'a', 0, 'found', 'a'],
|
||||||
|
['multiple words of text', 'uh-uh', 1],
|
||||||
|
['multiple words', 'multiple words, yeah', 0, 'found', 'multiple words'],
|
||||||
|
['\\(.*\\)c\\(.*\\)', 'abcde', 0, 'found+"-"+g1+"-"+g2', 'abcde-ab-de'],
|
||||||
|
['(\\(.*\\), \\(.*\\))', '(a, b)', 0, 'g2+"-"+g1', 'b-a'],
|
||||||
|
['[k]', 'ab', 1],
|
||||||
|
['a[-]?c', 'ac', 0, 'found', 'ac'],
|
||||||
|
['\\(abc\\)\\1', 'abcabc', 0, 'g1', 'abc'],
|
||||||
|
['\\([a-c]*\\)\\1', 'abcabc', 0, 'g1', 'abc'],
|
||||||
|
['^\\(.+\\)?B', 'AB', 0, 'g1', 'A'],
|
||||||
|
['\\(a+\\).\\1$', 'aaaaa', 0, 'found+"-"+g1', 'aaaaa-aa'],
|
||||||
|
['^\\(a+\\).\\1$', 'aaaa', 1],
|
||||||
|
['\\(abc\\)\\1', 'abcabc', 0, 'found+"-"+g1', 'abcabc-abc'],
|
||||||
|
['\\([a-c]+\\)\\1', 'abcabc', 0, 'found+"-"+g1', 'abcabc-abc'],
|
||||||
|
['\\(a\\)\\1', 'aa', 0, 'found+"-"+g1', 'aa-a'],
|
||||||
|
['\\(a+\\)\\1', 'aa', 0, 'found+"-"+g1', 'aa-a'],
|
||||||
|
['\\(a+\\)+\\1', 'aa', 0, 'found+"-"+g1', 'aa-a'],
|
||||||
|
['\\(a\\).+\\1', 'aba', 0, 'found+"-"+g1', 'aba-a'],
|
||||||
|
['\\(a\\)ba*\\1', 'aba', 0, 'found+"-"+g1', 'aba-a'],
|
||||||
|
['\\(aa\\|a\\)a\\1$', 'aaa', 0, 'found+"-"+g1', 'aaa-a'],
|
||||||
|
['\\(a\\|aa\\)a\\1$', 'aaa', 0, 'found+"-"+g1', 'aaa-a'],
|
||||||
|
['\\(a+\\)a\\1$', 'aaa', 0, 'found+"-"+g1', 'aaa-a'],
|
||||||
|
['\\([abc]*\\)\\1', 'abcabc', 0, 'found+"-"+g1', 'abcabc-abc'],
|
||||||
|
['\\(a\\)\\(b\\)c\\|ab', 'ab', 0, 'found+"-"+g1+"-"+g2', 'ab-None-None'],
|
||||||
|
['\\(a\\)+x', 'aaax', 0, 'found+"-"+g1', 'aaax-a'],
|
||||||
|
['\\([ac]\\)+x', 'aacx', 0, 'found+"-"+g1', 'aacx-c'],
|
||||||
|
['\\([^/]*/\\)*sub1/', 'd:msgs/tdir/sub1/trial/away.cpp', 0, 'found+"-"+g1', 'd:msgs/tdir/sub1/-tdir/'],
|
||||||
|
['\\([^.]*\\)\\.\\([^:]*\\):[T ]+\\(.*\\)', 'track1.title:TBlah blah blah', 0, 'found+"-"+g1+"-"+g2+"-"+g3', 'track1.title:TBlah blah blah-track1-title-Blah blah blah'],
|
||||||
|
['\\([^N]*N\\)+', 'abNNxyzN', 0, 'found+"-"+g1', 'abNNxyzN-xyzN'],
|
||||||
|
['\\([^N]*N\\)+', 'abNNxyz', 0, 'found+"-"+g1', 'abNN-N'],
|
||||||
|
['\\([abc]*\\)x', 'abcx', 0, 'found+"-"+g1', 'abcx-abc'],
|
||||||
|
['\\([abc]*\\)x', 'abc', 1],
|
||||||
|
['\\([xyz]*\\)x', 'abcx', 0, 'found+"-"+g1', 'x-'],
|
||||||
|
['\\(a\\)+b\\|aac', 'aac', 0, 'found+"-"+g1', 'aac-None'],
|
||||||
|
['\\<a', 'a', 0, 'found', 'a'],
|
||||||
|
['\\<a', '!', 1],
|
||||||
|
['a\\<b', 'ab', 1],
|
||||||
|
['a\\>', 'ab', 1],
|
||||||
|
['a\\>', 'a!', 0, 'found', 'a'],
|
||||||
|
['a\\>', 'a', 0, 'found', 'a'],
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
# test suite for match
|
||||||
|
match_regex_tests=[
|
||||||
|
['abc', 'abc', 0, 'found', 'abc'],
|
||||||
|
['abc', 'xbc', 1],
|
||||||
|
['abc', 'axc', 1],
|
||||||
|
['abc', 'abx', 1],
|
||||||
|
['abc', 'xabcy', 1],
|
||||||
|
['abc', 'ababc', 1],
|
||||||
|
['ab*c', 'abc', 0, 'found', 'abc'],
|
||||||
|
['ab*bc', 'abc', 0, 'found', 'abc'],
|
||||||
|
['ab*bc', 'abbc', 0, 'found', 'abbc'],
|
||||||
|
['ab*bc', 'abbbbc', 0, 'found', 'abbbbc'],
|
||||||
|
['ab+bc', 'abbc', 0, 'found', 'abbc'],
|
||||||
|
['ab+bc', 'abc', 1],
|
||||||
|
['ab+bc', 'abq', 1],
|
||||||
|
['ab+bc', 'abbbbc', 0, 'found', 'abbbbc'],
|
||||||
|
['ab?bc', 'abbc', 0, 'found', 'abbc'],
|
||||||
|
['ab?bc', 'abc', 0, 'found', 'abc'],
|
||||||
|
['ab?bc', 'abbbbc', 1],
|
||||||
|
['ab?c', 'abc', 0, 'found', 'abc'],
|
||||||
|
['^abc$', 'abc', 0, 'found', 'abc'],
|
||||||
|
['^abc$', 'abcc', 1],
|
||||||
|
['^abc', 'abcc', 0, 'found', 'abc'],
|
||||||
|
['^abc$', 'aabc', 1],
|
||||||
|
['abc$', 'aabc', 1],
|
||||||
|
['^', 'abc', 0, 'found+"-"', '-'],
|
||||||
|
['$', 'abc', 1],
|
||||||
|
['a.c', 'abc', 0, 'found', 'abc'],
|
||||||
|
['a.c', 'axc', 0, 'found', 'axc'],
|
||||||
|
['a.*c', 'axyzc', 0, 'found', 'axyzc'],
|
||||||
|
['a.*c', 'axyzd', 1],
|
||||||
|
['a[bc]d', 'abc', 1],
|
||||||
|
['a[bc]d', 'abd', 0, 'found', 'abd'],
|
||||||
|
['a[b-d]e', 'abd', 1],
|
||||||
|
['a[b-d]e', 'ace', 0, 'found', 'ace'],
|
||||||
|
['a[b-d]', 'aac', 1],
|
||||||
|
['a[-b]', 'a-', 0, 'found', 'a-'],
|
||||||
|
['a[b-]', 'a-', 0, 'found', 'a-'],
|
||||||
|
['a[]b', '-', 2],
|
||||||
|
['a[', '-', 2],
|
||||||
|
['a\\', '-', 2],
|
||||||
|
['abc\\)', '-', 2],
|
||||||
|
['\\(abc', '-', 2],
|
||||||
|
['a]', 'a]', 0, 'found', 'a]'],
|
||||||
|
['a[]]b', 'a]b', 0, 'found', 'a]b'],
|
||||||
|
['a[^bc]d', 'aed', 0, 'found', 'aed'],
|
||||||
|
['a[^bc]d', 'abd', 1],
|
||||||
|
['a[^-b]c', 'adc', 0, 'found', 'adc'],
|
||||||
|
['a[^-b]c', 'a-c', 1],
|
||||||
|
['a[^]b]c', 'a]c', 1],
|
||||||
|
['a[^]b]c', 'adc', 0, 'found', 'adc'],
|
||||||
|
['\\ba\\b', 'a-', 0, '"-"', '-'],
|
||||||
|
['\\ba\\b', '-a', 1],
|
||||||
|
['\\ba\\b', '-a-', 1],
|
||||||
|
['\\by\\b', 'xy', 1],
|
||||||
|
['\\by\\b', 'yz', 1],
|
||||||
|
['\\by\\b', 'xyz', 1],
|
||||||
|
['ab\\|cd', 'abc', 0, 'found', 'ab'],
|
||||||
|
['ab\\|cd', 'abcd', 0, 'found', 'ab'],
|
||||||
|
['\\(\\)ef', 'def', 1],
|
||||||
|
['$b', 'b', 1],
|
||||||
|
['a(b', 'a(b', 0, 'found+"-"+g1', 'a(b-None'],
|
||||||
|
['a(*b', 'ab', 0, 'found', 'ab'],
|
||||||
|
['a(*b', 'a((b', 0, 'found', 'a((b'],
|
||||||
|
['a\\\\b', 'a\\b', 0, 'found', 'a\\b'],
|
||||||
|
['\\(\\(a\\)\\)', 'abc', 0, 'found+"-"+g1+"-"+g2', 'a-a-a'],
|
||||||
|
['\\(a\\)b\\(c\\)', 'abc', 0, 'found+"-"+g1+"-"+g2', 'abc-a-c'],
|
||||||
|
['a+b+c', 'aabbabc', 1],
|
||||||
|
['\\(a+\\|b\\)*', 'ab', 0, 'found+"-"+g1', 'ab-b'],
|
||||||
|
['\\(a+\\|b\\)+', 'ab', 0, 'found+"-"+g1', 'ab-b'],
|
||||||
|
['\\(a+\\|b\\)?', 'ab', 0, 'found+"-"+g1', 'a-a'],
|
||||||
|
['\\)\\(', '-', 2],
|
||||||
|
['[^ab]*', 'cde', 0, 'found', 'cde'],
|
||||||
|
['abc', '', 1],
|
||||||
|
['a*', '', 0, 'found', ''],
|
||||||
|
['a\\|b\\|c\\|d\\|e', 'e', 0, 'found', 'e'],
|
||||||
|
['\\(a\\|b\\|c\\|d\\|e\\)f', 'ef', 0, 'found+"-"+g1', 'ef-e'],
|
||||||
|
['abcd*efg', 'abcdefg', 0, 'found', 'abcdefg'],
|
||||||
|
['ab*', 'xabyabbbz', 1],
|
||||||
|
['ab*', 'xayabbbz', 1],
|
||||||
|
['\\(ab\\|cd\\)e', 'abcde', 1],
|
||||||
|
['[abhgefdc]ij', 'hij', 0, 'found', 'hij'],
|
||||||
|
['^\\(ab\\|cd\\)e', 'abcde', 1, 'xg1y', 'xy'],
|
||||||
|
['\\(abc\\|\\)ef', 'abcdef', 1],
|
||||||
|
['\\(a\\|b\\)c*d', 'abcd', 1],
|
||||||
|
['\\(ab\\|ab*\\)bc', 'abc', 0, 'found+"-"+g1', 'abc-a'],
|
||||||
|
['a\\([bc]*\\)c*', 'abc', 0, 'found+"-"+g1', 'abc-bc'],
|
||||||
|
['a\\([bc]*\\)\\(c*d\\)', 'abcd', 0, 'found+"-"+g1+"-"+g2', 'abcd-bc-d'],
|
||||||
|
['a\\([bc]+\\)\\(c*d\\)', 'abcd', 0, 'found+"-"+g1+"-"+g2', 'abcd-bc-d'],
|
||||||
|
['a\\([bc]*\\)\\(c+d\\)', 'abcd', 0, 'found+"-"+g1+"-"+g2', 'abcd-b-cd'],
|
||||||
|
['a[bcd]*dcdcde', 'adcdcde', 0, 'found', 'adcdcde'],
|
||||||
|
['a[bcd]+dcdcde', 'adcdcde', 1],
|
||||||
|
['\\(ab\\|a\\)b*c', 'abc', 0, 'found+"-"+g1', 'abc-ab'],
|
||||||
|
['\\(\\(a\\)\\(b\\)c\\)\\(d\\)', 'abcd', 0, 'g1+"-"+g2+"-"+g3+"-"+g4', 'abc-a-b-d'],
|
||||||
|
['[a-zA-Z_][a-zA-Z0-9_]*', 'alpha', 0, 'found', 'alpha'],
|
||||||
|
['^a\\(bc+\\|b[eh]\\)g\\|.h$', 'abh', 1],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'effgz', 0, 'found+"-"+g1+"-"+g2', 'effgz-effgz-None'],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'ij', 0, 'found+"-"+g1+"-"+g2', 'ij-ij-j'],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'effg', 1],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'bcdd', 1],
|
||||||
|
['\\(bc+d$\\|ef*g.\\|h?i\\(j\\|k\\)\\)', 'reffgz', 1],
|
||||||
|
['\\(\\(\\(\\(\\(\\(\\(\\(\\(a\\)\\)\\)\\)\\)\\)\\)\\)\\)', 'a', 0, 'found', 'a'],
|
||||||
|
['multiple words of text', 'uh-uh', 1],
|
||||||
|
['multiple words', 'multiple words, yeah', 0, 'found', 'multiple words'],
|
||||||
|
['\\(.*\\)c\\(.*\\)', 'abcde', 0, 'found+"-"+g1+"-"+g2', 'abcde-ab-de'],
|
||||||
|
['(\\(.*\\), \\(.*\\))', '(a, b)', 0, 'g2+"-"+g1', 'b-a'],
|
||||||
|
['[k]', 'ab', 1],
|
||||||
|
['a[-]?c', 'ac', 0, 'found', 'ac'],
|
||||||
|
['\\(abc\\)\\1', 'abcabc', 0, 'g1', 'abc'],
|
||||||
|
['\\([a-c]*\\)\\1', 'abcabc', 0, 'g1', 'abc'],
|
||||||
|
['^\\(.+\\)?B', 'AB', 0, 'g1', 'A'],
|
||||||
|
['\\(a+\\).\\1$', 'aaaaa', 0, 'found+"-"+g1', 'aaaaa-aa'],
|
||||||
|
['^\\(a+\\).\\1$', 'aaaa', 1],
|
||||||
|
['\\(abc\\)\\1', 'abcabc', 0, 'found+"-"+g1', 'abcabc-abc'],
|
||||||
|
['\\([a-c]+\\)\\1', 'abcabc', 0, 'found+"-"+g1', 'abcabc-abc'],
|
||||||
|
['\\(a\\)\\1', 'aa', 0, 'found+"-"+g1', 'aa-a'],
|
||||||
|
['\\(a+\\)\\1', 'aa', 0, 'found+"-"+g1', 'aa-a'],
|
||||||
|
['\\(a+\\)+\\1', 'aa', 0, 'found+"-"+g1', 'aa-a'],
|
||||||
|
['\\(a\\).+\\1', 'aba', 0, 'found+"-"+g1', 'aba-a'],
|
||||||
|
['\\(a\\)ba*\\1', 'aba', 0, 'found+"-"+g1', 'aba-a'],
|
||||||
|
['\\(aa\\|a\\)a\\1$', 'aaa', 0, 'found+"-"+g1', 'aaa-a'],
|
||||||
|
['\\(a\\|aa\\)a\\1$', 'aaa', 0, 'found+"-"+g1', 'aaa-a'],
|
||||||
|
['\\(a+\\)a\\1$', 'aaa', 0, 'found+"-"+g1', 'aaa-a'],
|
||||||
|
['\\([abc]*\\)\\1', 'abcabc', 0, 'found+"-"+g1', 'abcabc-abc'],
|
||||||
|
['\\(a\\)\\(b\\)c\\|ab', 'ab', 0, 'found+"-"+g1+"-"+g2', 'ab-None-None'],
|
||||||
|
['\\(a\\)+x', 'aaax', 0, 'found+"-"+g1', 'aaax-a'],
|
||||||
|
['\\([ac]\\)+x', 'aacx', 0, 'found+"-"+g1', 'aacx-c'],
|
||||||
|
['\\([^/]*/\\)*sub1/', 'd:msgs/tdir/sub1/trial/away.cpp', 0, 'found+"-"+g1', 'd:msgs/tdir/sub1/-tdir/'],
|
||||||
|
['\\([^.]*\\)\\.\\([^:]*\\):[T ]+\\(.*\\)', 'track1.title:TBlah blah blah', 0, 'found+"-"+g1+"-"+g2+"-"+g3', 'track1.title:TBlah blah blah-track1-title-Blah blah blah'],
|
||||||
|
['\\([^N]*N\\)+', 'abNNxyzN', 0, 'found+"-"+g1', 'abNNxyzN-xyzN'],
|
||||||
|
['\\([^N]*N\\)+', 'abNNxyz', 0, 'found+"-"+g1', 'abNN-N'],
|
||||||
|
['\\([abc]*\\)x', 'abcx', 0, 'found+"-"+g1', 'abcx-abc'],
|
||||||
|
['\\([abc]*\\)x', 'abc', 1],
|
||||||
|
['\\([xyz]*\\)x', 'abcx', 1],
|
||||||
|
['\\(a\\)+b\\|aac', 'aac', 0, 'found+"-"+g1', 'aac-None'],
|
||||||
|
['\\<a', 'a', 0, 'found', 'a'],
|
||||||
|
['\\<a', '!', 1],
|
||||||
|
['a\\<b', 'ab', 1],
|
||||||
|
['a\\>', 'ab', 1],
|
||||||
|
['a\\>', 'a!', 0, 'found', 'a'],
|
||||||
|
['a\\>', 'a', 0, 'found', 'a'],
|
||||||
|
]
|
||||||
|
|
||||||
|
# test suite for split()
|
||||||
|
# element 0: pattern
|
||||||
|
# 1: string to split
|
||||||
|
# 3: compile result
|
||||||
|
# 4: maxsplit
|
||||||
|
# 5: splitted fields list
|
||||||
|
split_regex_tests = [
|
||||||
|
["[ |,]", "with you, nothing, and me", 0, 0, ["with","you","nothing","and","me"]],
|
||||||
|
["[ |,]", "with you, nothing, and me", 0, 1, ["with", "you, nothing, and me"]],
|
||||||
|
["\\ ", "send email to apply", 0, 0, ["send", "email", "to", "apply"]],
|
||||||
|
["\\ ", "send email to apply", 0, 2, ["send", "email", "to apply"]],
|
||||||
|
["[+ | -]", "+86-028-83201034", 0, 0, ["86", "028", "83201034"]],
|
||||||
|
["[+ | -]", "+86-028-83201034", 0, 1, ["86", "028-83201034"]],
|
||||||
|
["[*|#]", "slide show", 0, 0, ["slide show"]],
|
||||||
|
["(", "whats ever", 0, 1, ["whats ever"]],
|
||||||
|
["@#!~$%^&*()<>\n", "who knows", 0, 1, ["who knows"]],
|
||||||
|
]
|
||||||
|
|
||||||
|
# test suite for findall()
|
||||||
|
# element 0: pattern
|
||||||
|
# 1: string to match
|
||||||
|
# 3: compile result
|
||||||
|
# 4: starting position
|
||||||
|
# 5: grouped fields list
|
||||||
|
|
||||||
|
# reobj.find()
|
||||||
|
findall_regex_tests = [
|
||||||
|
["\\ ", "send email to apply", 0, 0, [" ", " ", " "]],
|
||||||
|
["\\ ", "send email to apply", 0, 5, [" ", " "]],
|
||||||
|
["[+ | -]", "+86-028-83201034", 0, 0, ["+", "-", "-"]],
|
||||||
|
["[+ | -]", "+86-028-83201034", 0, 1, ["-", "-"]],
|
||||||
|
["sl.*e\\|#", "slide show at Room #3", 0, 0, ["slide", "#"]],
|
||||||
|
["w.+s\\|e.*r", "whats ever", 0, 0, ["whats", "ever"]],
|
||||||
|
["Euler\\|Gauss", "Both Euler and Gauss are great mathematicians", 0, 0, ["Euler", "Gauss"]],
|
||||||
|
]
|
||||||
|
|
||||||
|
# module re.findall()
|
||||||
|
mod_findall_regex_tests = [
|
||||||
|
["\\ ", "send email to apply", 0, 0, [" ", " ", " "]],
|
||||||
|
["\\ ", "send email to apply", 0, 0, [" ", " ", " "]],
|
||||||
|
["[+ | -]", "+86-028-83201034", 0, 0, ["+", "-", "-"]],
|
||||||
|
["[+ | -]", "+86-028-83201034", 0, 0, ["+", "-", "-"]],
|
||||||
|
["sl.*e\\|#", "slide show at Room #3", 0, 0, ["slide", "#"]],
|
||||||
|
["w.+s\\|e.*r", "whats ever", 0, 0, ["whats", "ever"]],
|
||||||
|
["Euler\\|Gauss", "Both Euler and Gauss are great mathematicians", 0, 0, ["Euler", "Gauss"]],
|
||||||
|
]
|
||||||
|
|
||||||
|
# test for match object's groups() method
|
||||||
|
# element 0: pattern
|
||||||
|
# 1: string
|
||||||
|
# 2: compile result
|
||||||
|
# 3: matched fields, for groups()
|
||||||
|
# 4: group index, valid when > 0, for start(), end(), and span()
|
||||||
|
# 5: pattern's starting index in string, for start() and span()
|
||||||
|
# 6: pattern's ending index in string, for end() and span
|
||||||
|
matobj_groups_regex_tests = [
|
||||||
|
["\\(abc\\(.*xyz\\)\\(.*31415926\\)\\)", "where is abc and flurry xyz, which is pi 31415926, derived from ms", 0, ["abc and flurry xyz, which is pi 31415926"," and flurry xyz",", which is pi 31415926"], 2, 12, 27],
|
||||||
|
|
||||||
|
["[a\\|b]\\(.+\\)shoe\\([t]+\\)d", "bbbshoetttdxrznmlkjp", 0, ["bb", "ttt"], 1, 1, 3],
|
||||||
|
|
||||||
|
["abcdef", "xyah2oewoyqe030uabcdefwhalsdewnkhgiohyczb", 0, [], -1, 0, 0],
|
||||||
|
]
|
||||||
|
|
Reference in New Issue
Block a user