Why Gemfury? Push, build, and install  RubyGems npm packages Python packages Maven artifacts PHP packages Go Modules Debian packages RPM packages NuGet packages

Repository URL to install this package:

Details    
Size: Mime:
ó
¡‹EYc@swdZddlZddlZddlmZddlmZddlmZddl	m
Z
ddlmZm
Z
mZmZe
dƒZejd	ejƒZejd
ejƒZejdƒZd„ZeƒZ[ejd
ƒZejdƒZedƒZedƒZedƒZedƒZedƒZedƒZ edƒZ!edƒZ"edƒZ#edƒZ$edƒZ%edƒZ&edƒZ'edƒZ(edƒZ)edƒZ*edƒZ+ed ƒZ,ed!ƒZ-ed"ƒZ.ed#ƒZ/ed$ƒZ0ed%ƒZ1ed&ƒZ2ed'ƒZ3ed(ƒZ4ed)ƒZ5ed*ƒZ6ed+ƒZ7ed,ƒZ8ed-ƒZ9ed.ƒZ:ed/ƒZ;ed0ƒZ<ed1ƒZ=ed2ƒZ>ed3ƒZ?ed4ƒZ@ed5ƒZAed6ƒZBed7ƒZCed8ƒZDed9ƒZEed:ƒZFed;ƒZGed<ƒZHed=ƒZIed>ƒZJed?ƒZKied@6e3dA6edB6e"dC6e+dD6e*dE6e.dF6e4dG6e&dH6e0dI6e'dJ6e1dK6e%dL6e/dM6e!dN6e,dO6e#dP6e$dQ6e(dR6e)dS6edT6e dU6edV6e-dW6edX6e2dY6ZLeMgeeLƒD]\ZNZOeOeNf^qƒZPeQeLƒeQePƒksUtRdZƒ‚ejd[dWjSd\„eTeLd]d^„ƒDƒƒƒZUeVeAeCeBe5eFeGeHgƒZWeVe5eIeCeHgƒZXd_„ZYd`„ZZda„Z[db„Z\dc„Z]dde^fde„ƒYZ_dfe`fdg„ƒYZae
dhe^fdi„ƒYƒZbe
dje^fdk„ƒYƒZcdl„Zddme^fdn„ƒYZedS(osã
    jinja2.lexer
    ~~~~~~~~~~~~

    This module implements a Jinja / Python combination lexer. The
    `Lexer` class provided by this module is used to do some preprocessing
    for Jinja.

    On the one hand it filters out invalid operators like the bitshift
    operators we don't allow in templates. On the other hand it separates
    template code and python code in expressions.

    :copyright: (c) 2017 by the Jinja Team.
    :license: BSD, see LICENSE for more details.
iÿÿÿÿN(t
itemgetter(tdeque(tTemplateSyntaxError(tLRUCache(t	iteritemstimplements_iteratort	text_typetinterni2s\s+s7('([^'\\]*(?:\\.[^'\\]*)*)'|"([^"\\]*(?:\\.[^"\\]*)*)")s\d+cCsŠytdddƒWntk
r1tjdƒSXddl}ddlm}tjd|j|jfƒ}tjj	dƒ~|`|S(	Nsföös	<unknown>tevals\b[a-zA-Z_][a-zA-Z0-9_]*\biÿÿÿÿ(t_stringdefss	[%s][%s]*sjinja2._stringdefs(
tcompiletSyntaxErrortretjinja2R	t	xid_starttxid_continuetsystmodulestpop(R
R	tname_re((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt
_make_name_re%s
s(?<!\.)\d+\.\d+s(\r\n|\r|\n)taddtassigntcolontcommatdivtdotteqtfloordivtgttgteqtlbracetlbrackettlparentlttlteqtmodtmultnetpipetpowtrbracetrbrackettrparent	semicolontsubttildet
whitespacetfloattintegertnametstringtoperatortblock_begint	block_endtvariable_begintvariable_endt	raw_begintraw_endt
comment_begintcomment_endtcommenttlinestatement_begintlinestatement_endtlinecomment_begintlinecomment_endtlinecommenttdatatinitialteoft+t-t/s//t*t%s**t~t[t]t(t)t{t}s==s!=t>s>=t<s<=t=t.t:t|t,t;soperators droppeds(%s)ccs|]}tj|ƒVqdS(N(Rtescape(t.0tx((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pys	<genexpr>’stkeycCst|ƒS(N(tlen(R\((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt<lambda>“scCsx|tkrt|Sidt6dt6dt6dt6dt6dt6dt6dt6dt	6d	t
6d
t6dt6j
||ƒS(Nsbegin of commentsend of commentR=sbegin of statement blocksend of statement blocksbegin of print statementsend of print statementsbegin of line statementsend of line statementstemplate data / textsend of template(treverse_operatorstTOKEN_COMMENT_BEGINtTOKEN_COMMENT_ENDt
TOKEN_COMMENTtTOKEN_LINECOMMENTtTOKEN_BLOCK_BEGINtTOKEN_BLOCK_ENDtTOKEN_VARIABLE_BEGINtTOKEN_VARIABLE_ENDtTOKEN_LINESTATEMENT_BEGINtTOKEN_LINESTATEMENT_ENDt
TOKEN_DATAt	TOKEN_EOFtget(t
token_type((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt_describe_token_types 
cCs#|jdkr|jSt|jƒS(s#Returns a description of the token.R2(ttypetvalueRo(ttoken((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pytdescribe_token°scCsGd|kr7|jddƒ\}}|dkr=|Sn|}t|ƒS(s0Like `describe_token` but for token expressions.RViR2(tsplitRo(texprRpRq((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pytdescribe_token_expr·scCsttj|ƒƒS(ssCount the number of newline characters in the string.  This is
    useful for extensions that filter a stream.
    (R^t
newline_retfindall(Rq((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pytcount_newlinesÂscCstj}t|jƒd||jƒft|jƒd||jƒft|jƒd||jƒfg}|jd
k	r§|jt|jƒdd||jƒfƒn|j	d
k	rå|jt|j	ƒdd||j	ƒfƒngt
|dtƒD]}|d	^qøS(sACompiles all the rules from the environment into a list of rules.R=tblocktvariablet
linestatements	^[ \t\v]*RBs(?:^|(?<=\S))[^\S\r\n]*treverseiN(RRZR^tcomment_start_stringtblock_start_stringtvariable_start_stringtline_statement_prefixtNonetappendtline_comment_prefixtsortedtTrue(tenvironmenttetrulesR\((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt
compile_rulesÉs	tFailurecBs#eZdZed„Zd„ZRS(sjClass that raises a `TemplateSyntaxError` if called.
    Used by the `Lexer` to specify known errors.
    cCs||_||_dS(N(tmessageterror_class(tselfRŒtcls((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt__init__ås	cCs|j|j||ƒ‚dS(N(RRŒ(RŽtlinenotfilename((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt__call__és(t__name__t
__module__t__doc__RRR“(((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR‹àstTokencBs`eZdZdZd„edƒDƒ\ZZZd„Zd„Z	d„Z
d„Zd„ZRS(	sToken class.ccs!|]}tt|ƒƒVqdS(N(tpropertyR(R[R\((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pys	<genexpr>ðsicCs%tj||tt|ƒƒ|fƒS(N(ttuplet__new__Rtstr(RR‘RpRq((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyRšòscCs7|jtkrt|jS|jdkr0|jS|jS(NR2(RpR`Rq(RŽ((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt__str__õs
cCsE|j|krtSd|krA|jddƒ|j|jgkStS(s¸Test a token against a token expression.  This can either be a
        token type or ``'token_type:token_value'``.  This can only test
        against string values and types.
        RVi(RpR†RtRqtFalse(RŽRu((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyttestüs
"cGs(x!|D]}|j|ƒrtSqWtS(s(Test against multiple token expressions.(RžR†R(RŽtiterableRu((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyttest_any	s
cCsd|j|j|jfS(NsToken(%r, %r, %r)(R‘RpRq(RŽ((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt__repr__s((
R”R•R–t	__slots__trangeR‘RpRqRšRœRžR R¡(((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR—ís			
	tTokenStreamIteratorcBs)eZdZd„Zd„Zd„ZRS(s`The iterator for tokenstreams.  Iterate over the stream
    until the eof token is reached.
    cCs
||_dS(N(tstream(RŽR¥((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyRscCs|S(N((RŽ((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt__iter__!scCsE|jj}|jtkr4|jjƒtƒ‚nt|jƒ|S(N(R¥tcurrentRpRltcloset
StopIterationtnext(RŽRr((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt__next__$s

(R”R•R–RR¦R«(((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR¤s		tTokenStreamcBseZdZd„Zd„Zd„ZeZed„ddƒZd„Z	d„Z
d	d
„Zd„Zd„Z
d
„Zd„Zd„ZRS(sÛA token stream is an iterable that yields :class:`Token`\s.  The
    parser however does not iterate over it but calls :meth:`next` to go
    one token ahead.  The current active token is stored as :attr:`current`.
    cCsYt|ƒ|_tƒ|_||_||_t|_tdt	dƒ|_
t|ƒdS(Nit(titert_iterRt_pushedR2R’RtclosedR—t
TOKEN_INITIALR§Rª(RŽt	generatorR2R’((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR4s			cCs
t|ƒS(N(R¤(RŽ((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR¦=scCst|jƒp|jjtk	S(N(tboolR°R§RpRl(RŽ((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt__bool__@scCs|S(N((R\((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR_Dstdocs Are we at the end of the stream?cCs|jj|ƒdS(s Push a token back to the stream.N(R°Rƒ(RŽRr((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pytpushFscCs/t|ƒ}|j}|j|ƒ||_|S(sLook at the next token.(RªR§R·(RŽt	old_tokentresult((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pytlookJs
	
	icCs%xt|ƒD]}t|ƒq
WdS(sGot n tokens ahead.N(R£Rª(RŽtnR\((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pytskipRscCs |jj|ƒrt|ƒSdS(sqPerform the token test and return the token if it matched.
        Otherwise the return value is `None`.
        N(R§RžRª(RŽRu((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pytnext_ifWscCs|j|ƒdk	S(s8Like :meth:`next_if` but only returns `True` or `False`.N(R½R‚(RŽRu((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pytskip_if^scCst|j}|jr'|jjƒ|_nI|jjtk	rpyt|jƒ|_Wqptk
rl|jƒqpXn|S(s)Go one token ahead and return the old one(	R§R°tpopleftRpRlRªR¯R©R¨(RŽtrv((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR«bs		
cCs1t|jjtdƒ|_d|_t|_dS(sClose the stream.R­N(R—R§R‘RlR‚R¯R†R±(RŽ((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR¨ns	cCs¬|jj|ƒst|ƒ}|jjtkrXtd||jj|j|jƒ‚ntd|t	|jƒf|jj|j|jƒ‚nz|jSWdt
|ƒXdS(s}Expect a given token type and return it.  This accepts the same
        argument as :meth:`jinja2.lexer.Token.test`.
        s(unexpected end of template, expected %r.sexpected token %r, got %rN(R§RžRvRpRlRR‘R2R’RsRª(RŽRu((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pytexpectts		(R”R•R–RR¦Rµt__nonzero__R˜teosR·RºR¼R½R¾R«R¨RÁ(((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR¬-s										cCs†|j|j|j|j|j|j|j|j|j|j	|j
|jf}tj
|ƒ}|dkr‚t|ƒ}|t|<n|S(s(Return a lexer which is probably cached.N(Rtblock_end_stringR€tvariable_end_stringR~tcomment_end_stringRR„ttrim_blockst
lstrip_blockstnewline_sequencetkeep_trailing_newlinet_lexer_cacheRmR‚tLexer(R‡R]tlexer((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt	get_lexer‰s"
RÌcBsPeZdZd„Zd„Zdddd„Zddd„Zddd„ZRS(s
Class that implements a lexer for a given environment. Automatically
    created by the environment class, usually you don't have to do that.

    Note that the lexer is not automatically bound to an environment.
    Multiple environments can share the same lexer.
    cCs4d„}tj}ttdfttdfttdft	t
dfttdft
tdfg}t|ƒ}|jr{dp~d}i}|jrè|dƒ}|d||jƒƒ}	|	j|jƒ}
||
réd||
jdƒƒpìd7}|	j|jƒ}
||
r$d||
jdƒƒp'd7}|d||jƒƒ}|j|jƒ}
|
rud||
jdƒƒpxd}d	}
d
|
||jƒ|||jƒf}d|
||jƒ|||jƒf}||d<||d
<nd||jƒ}|j|_|j|_i|ddjd||jƒ|||jƒ||jƒfgg|D]+\}}d|||j||ƒf^qZƒƒtdfdf|dƒtdfgd6|d||jƒ||jƒ|fƒttfdf|dƒt dƒfdfgt!6|d||jƒ||jƒ|fƒt"dfg|t#6|d||j$ƒ||j$ƒfƒt%dfg|t&6|d||jƒ|||jƒ||jƒ|fƒtt'fdf|dƒt dƒfdfgt(6|dƒt)dfg|t*6|dƒt+t,fdfgt-6|_.dS( NcSstj|tjtjBƒS(N(RR
tMtS(R\((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR_¨ss\n?R­RFs^%s(.*)s|%sis(?!%s)s^[ \t]*s%s%s(?!%s)|%s\+?s%s%s%s|%s\+?RzR=s%ss(.*?)(?:%s)RWs4(?P<raw_begin>(?:\s*%s\-|%s)\s*raw\s*(?:\-%s\s*|%s))s(?P<%s_begin>\s*%s\-|%s)s#bygroups.+troots(.*?)((?:\-%s\s*|%s)%s)s#pops(.)sMissing end of comment tags(?:\-%s\s*|%s)%ss
\-%s\s*|%ss1(.*?)((?:\s*%s\-|%s)\s*endraw\s*(?:\-%s\s*|%s%s))sMissing end of raw directives	\s*(\n|$)s(.*?)()(?=\n|$)(/RRZt
whitespace_retTOKEN_WHITESPACER‚tfloat_retTOKEN_FLOATt
integer_ret
TOKEN_INTEGERRt
TOKEN_NAMEt	string_retTOKEN_STRINGtoperator_retTOKEN_OPERATORRŠRÇRÈRtmatchR~tgroupR€RÉRÊtjoinRÄRmRkRÆRcRbR‹RaRfReRÅRhRgt
TOKEN_RAW_ENDtTOKEN_RAW_BEGINRjRiRdtTOKEN_LINECOMMENT_ENDtTOKEN_LINECOMMENT_BEGINR‰(RŽR‡tcRˆt	tag_rulestroot_tag_rulestblock_suffix_ret	prefix_retno_lstrip_ret
block_difftmtcomment_difftno_variable_ret	lstrip_retblock_prefix_retcomment_prefix_reR»tr((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR¦sš				))%

:
"

"cCstj|j|ƒS(s@Called for strings and template data to normalize it to unicode.(RwR-RÉ(RŽRq((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt_normalize_newlines&scCs7|j||||ƒ}t|j|||ƒ||ƒS(sCCalls tokeniter + tokenize and wraps it in a token stream.
        (t	tokeniterR¬twrap(RŽtsourceR2R’tstateR¥((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyttokenize*sc	cs¢x›|D]“\}}}|tkr(qna|dkr=d}nL|dkrRd}n7|dkrdqn%|dkr‚|j|ƒ}n|dkr—|}nò|d	kr²t|ƒ}n×|d
kr:y/|j|dd!ƒjd
dƒjdƒ}Wq‰tk
r6}t|ƒjdƒdjƒ}t||||ƒ‚q‰XnO|dkrUt	|ƒ}n4|dkrpt
|ƒ}n|dkr‰t|}nt|||ƒVqWdS(s‹This is called with the stream as returned by `tokenize` and wraps
        every token in a :class:`Token` and converts the value.
        R>R5R?R6R9R:RCtkeywordR2R3iiÿÿÿÿtasciitbackslashreplacesunicode-escapeRVR1R0R4N(R9R:(
tignored_tokensRòR›tencodetdecodet	ExceptionRttstripRtintR0t	operatorsR—(	RŽR¥R2R’R‘RrRqRˆtmsg((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyRô0s<			

c	csrt|ƒ}|jƒ}|jr[|r[x1d D]&}|j|ƒr.|jdƒPq.q.Wndj|ƒ}d}d}dg}	|dk	rÃ|dkrÃ|d!ks¯td
ƒ‚|	j|dƒnd}|j|	d}
t	|ƒ}g}xxy|
D]>\}
}}|
j
||ƒ}|dkr)qön|rA|d"krAqönt|tƒrMx
t
|ƒD]é\}}|jtkrŠ|||ƒ‚q]|dkrýx­t|jƒƒD]=\}}|dk	r©|||fV||jdƒ7}Pq©q©Wtd|
ƒ‚q]|j|dƒ}|s"|tkr3|||fVn||jdƒ7}q]Wn|jƒ}|dkr'|dkr|jdƒq'|dkr|jdƒq'|dkr¹|jdƒq'|d#kr'|sçtd||||ƒ‚n|jƒ}||kr$td||f|||ƒ‚q$q'n|s9|tkrJ|||fVn||jdƒ7}|jƒ}|dk	r|dkrŽ|	jƒnl|dkríx]t|jƒƒD])\}}|dk	r­|	j|ƒPq­q­Wtd|
ƒ‚n
|	j|ƒ|j|	d}
n||kr-td|
ƒ‚n|}PqöW||krHdStd|||f|||ƒ‚qïdS($sThis method tokenizes the text and returns the tokens in a
        generator.  Use this method if you just want to tokenize a template.
        s
s
s
R­iiRÑR{Rzs
invalid statet_beginiÿÿÿÿR8R6R?s#bygroups?%r wanted to resolve the token dynamically but no group matchedR4RPRQRNRORLRMsunexpected '%s'sunexpected '%s', expected '%s's#popsC%r wanted to resolve the new state dynamically but no group matcheds,%r yielded empty string without stack changeNsunexpected char %r at %d(s
s
s
(svariablesblock(R8s	block_endslinestatement_end(RQRORM(Rt
splitlinesRÊtendswithRƒRßR‚tAssertionErrorR‰R^RÝt
isinstanceR™t	enumeratet	__class__R‹Rt	groupdicttcounttRuntimeErrorRÞtignore_if_emptyRRtend(RŽRõR2R’RötlinestnewlinetposR‘tstacktstatetokenst
source_lengthtbalancing_stacktregexttokenst	new_stateRëtidxRrR]RqRCtexpected_optpos2((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyRóUs°

		







N(	R”R•R–RRòR‚R÷RôRó(((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyR̞s	€	%(fR–RRR4RtcollectionsRtjinja2.exceptionsRtjinja2.utilsRtjinja2._compatRRRRRËR
tURÒRÐRÙRÖRRRÔRwt	TOKEN_ADDtTOKEN_ASSIGNtTOKEN_COLONtTOKEN_COMMAt	TOKEN_DIVt	TOKEN_DOTtTOKEN_EQtTOKEN_FLOORDIVtTOKEN_GTt
TOKEN_GTEQtTOKEN_LBRACEtTOKEN_LBRACKETtTOKEN_LPARENtTOKEN_LTt
TOKEN_LTEQt	TOKEN_MODt	TOKEN_MULtTOKEN_NEt
TOKEN_PIPEt	TOKEN_POWtTOKEN_RBRACEtTOKEN_RBRACKETtTOKEN_RPARENtTOKEN_SEMICOLONt	TOKEN_SUBtTOKEN_TILDERÓRÕR×RØRÚRÜReRfRgRhRáRàRaRbRcRiRjRãRâRdRkR²RlRtdicttktvR`R^RRßR…RÛt	frozensetRûR
RoRsRvRyRŠtobjectR‹R™R—R¤R¬RÎRÌ(((sA/home/tvault/.virtenv/lib/python2.7/site-packages/jinja2/lexer.pyt<module>sè"			
1$ 							
+[