!C99Shell v. 1.0 pre-release build #16!

Software: Apache/2.0.54 (Fedora). PHP/5.0.4 

uname -a: Linux mina-info.me 2.6.17-1.2142_FC4smp #1 SMP Tue Jul 11 22:57:02 EDT 2006 i686 

uid=48(apache) gid=48(apache) groups=48(apache)
context=system_u:system_r:httpd_sys_script_t
 

Safe-mode: OFF (not secure)

/usr/bin/X11/./../../lib/evolution-openldap/../python2.4/site-packages/Ft/Lib/DistExt/   drwxr-xr-x
Free 4.85 GB of 27.03 GB (17.95%)
Home    Back    Forward    UPDIR    Refresh    Search    Buffer    Encoder    Tools    Proc.    FTP brute    Sec.    SQL    PHP-code    Update    Feedback    Self remove    Logout    


Viewing file:     GenerateL10n.py (10.95 KB)      -rw-r--r--
Select action/file-type:
(+) | (+) | (+) | Code (+) | Session (+) | (+) | SDB (+) | (+) | (+) | (+) | (+) | (+) |
import os, time
from distutils.core import Command
from distutils.dep_util import newer_group

POT_HEADER = """\
# %(name)s LANGUAGE message catalog
# Copyright %(year)s %(author)s
# FIRST AUTHOR <EMAIL@ADDRESS>, YEAR.
#
msgid ""
msgstr ""
"Project-Id-Version: %(name)s %(version)s\\n"
"POT-Creation-Date: %(creation-date)s\\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\\n"
"Language-Team: LANGUAGE <LL@li.org>\\n"
"MIME-Version: 1.0\\n"
"Content-Type: text/plain; charset=CHARSET\\n"
"Content-Transfer-Encoding: ENCODING\\n"
"Generated-By: %(module)s %(version)s\\n"

"""

_escape_chars = None
def _c_escape(s):
    global _escape_chars
    if _escape_chars is None:
        _escape_chars = {'\0' : '\\0',
                         '\a' : '\\a',
                         '\b' : '\\b',
                         '\f' : '\\f',
                         '\n' : '\\n',
                         '\r' : '\\r',
                         '\t' : '\\t',
                         '\v' : '\\v',
                         '\"' : '\\"',
                         }
        for i in xrange(128, 256):
            _escape_chars[chr(i)] = '\\%03o' % i
        
    res = list(s)
    for i in xrange(len(res)):
        if _escape_chars.has_key(res[i]):
            res[i] = _escape_chars[res[i]]
    return ''.join(res)
    

class GenerateL10n(Command):

    command_name = 'generate_l10n'

    description = "extract translatable strings from source"

    user_options = [
        ('force', 'f',
         'force locale generatation (overwrite existing files)'),
        ]

    boolean_options = ['force']

    def initialize_options(self):
        self.force = None
        return

    def finalize_options(self):
        self.set_undefined_options('generate', ('force', 'force'))
        self.translations_dir = 'po'
        return

    def run(self):
        sources = self.get_sources()
        outfile = self.get_pot_filename()
        dependencies = [__file__] + sources
                       
        if not (self.force or newer_group(dependencies, outfile, 'newer')):
            self.announce("skipping catalog generation (up-to-date)", 1)
            return

        # Get the translatable strings from the source files
        messages = self.extract_messages(sources)

        # Write the message catalog (POT file)
        self.write_catalog(messages)
        return

    # -- worker functions ----------------------------------------------

    def extract_messages(self, filenames):
        extractor = Extractor()

        for filename in filenames:
            self.announce("extracting strings from %s" % filename, 1)
            extractor.process(filename)

        return extractor.get_messages()

    def write_catalog(self, messages):

        def normalize(s):
            # This converts the various Python string types into a format
            # that is appropriate for .po files.
            lines = s.split('\n')
            if len(lines) == 1:
                s = '"' + _c_escape(s) + '"'
            else:
                if not lines[-1]:
                    del lines[-1]
                    lines[-1] = lines[-1] + '\n'
                for i in range(len(lines)):
                    lines[i] = _c_escape(lines[i])
                lineterm = '\\n"\n"'
                s = '""\n"' + lineterm.join(lines) + '"'
            return s


        self.mkpath(self.translations_dir)
        outfile = self.get_pot_filename()
        self.announce("writing message catalog %s" % outfile, 2)

        if self.dry_run:
            return

        # Binary mode is needed when generating on non-LF terminating
        # platforms (i.e. Windows, Mac).  Python can always read LF
        # terminated files.
        fp = open(outfile, 'wb')
        try:
            timestamp = time.time()
            repl = {'name' : self.distribution.get_name(),
                    'version' : self.distribution.get_version(),
                    'author' : self.distribution.get_author(),
                    'year' : time.strftime('%Y', time.localtime(timestamp)),
                    'creation-date' : time.strftime('%Y-%m-%d %H:%M+0000',
                                                    time.gmtime(timestamp)),
                    'module' : __name__,
                    }
            print >> fp, POT_HEADER % repl

            for msgstr, msglocs in messages:
                isdocstring = 0
                # If the entry was gleaned out of a docstring, then add a
                # comment stating so.  This is to aid translators who may wish
                # to skip translating some unimportant docstrings.
                if reduce(operator.__add__, msglocs.values()):
                    isdocstring = 1

                # msglocs is a dictionary-set of (filename, lineno) tuples.
                # We want to sort the entries in msglocs first by filename
                # and then by line number.
                msglocs = msglocs.keys()
                msglocs.sort()

                # write each location on a separate line
                for filename, lineno in msglocs:
                    d = {'filename': filename, 'lineno': lineno}
                    print >> fp, '#: %(filename)s:%(lineno)d' % d

                if isdocstring:
                    print >> fp, '#, docstring'

                print >> fp, 'msgid', normalize(msgstr)
                print >> fp, 'msgstr ""\n'
        finally:
            fp.close()
        return

    # -- utility functions ---------------------------------------------

    def get_sources(self):
        sources = []

        # Gather the Python sources
        if self.distribution.has_pure_modules():
            cmd = self.get_finalized_command('build_py')
            sources.extend(cmd.get_source_files())

        # Don't scan third-party additions for translatable messages
        exclude = os.sep + 'ThirdParty' + os.sep
        sources = [ fn for fn in sources if fn.find(exclude) == -1 ]
        return sources

    def get_pot_filename(self):
        return os.path.join(self.translations_dir,
                            self.distribution.get_name() + '.pot')

    # -- external interfaces -------------------------------------------

    def get_outputs(self):
        if self.distribution.l10n:
            return [self.get_pot_filename()]
        return []

import token
import tokenize
import operator

class Extractor:

    def __init__(self, excludedStrings=None, keywords=None, docstrings=0):
        self.__excluded = excludedStrings or []
        self.__keywords = ['_']
        if keywords:
            self.__keywords.extend(keywords)
        self.__docstrings = docstrings

        self.__messages = {}
        self.__state = self.__waiting
        self.__data = []
        self.__lineno = -1
        self.__freshmodule = 1
        self.__curfile = None
        return

    def __call__(self, ttype, tstring, stup, etup, line):
        # dispatch
##        import token
##        print >> sys.stderr, 'ttype:', token.tok_name[ttype], \
##              'tstring:', tstring
        self.__state(ttype, tstring, stup[0])

    def __waiting(self, ttype, tstring, lineno):
        # Do docstring extractions, if enabled
        if self.__docstrings:
            # module docstring?
            if self.__freshmodule:
                if ttype == tokenize.STRING:
                    self.__addentry(eval(tstring), lineno, isdocstring=1)
                    self.__freshmodule = 0
                elif ttype not in (tokenize.COMMENT, tokenize.NL):
                    self.__freshmodule = 0
                return
            # class docstring?
            if ttype == tokenize.NAME and tstring in ('class', 'def'):
                self.__state = self.__suiteseen
                return
        if ttype == tokenize.NAME and tstring in self.__keywords:
            self.__state = self.__keywordseen

    def __suiteseen(self, ttype, tstring, lineno):
        # ignore anything until we see the colon
        if ttype == tokenize.OP and tstring == ':':
            self.__state = self.__suitedocstring

    def __suitedocstring(self, ttype, tstring, lineno):
        # ignore any intervening noise
        if ttype == tokenize.STRING:
            self.__addentry(eval(tstring), lineno, isdocstring=1)
            self.__state = self.__waiting
        elif ttype not in (tokenize.NEWLINE, tokenize.INDENT,
                           tokenize.COMMENT):
            # there was no class docstring
            self.__state = self.__waiting

    def __keywordseen(self, ttype, tstring, lineno):
        if ttype == tokenize.OP and tstring == '(':
            self.__data = []
            self.__lineno = lineno
            self.__state = self.__openseen
        else:
            self.__state = self.__waiting

    def __openseen(self, ttype, tstring, lineno):
        if ttype == tokenize.OP and tstring == ')':
            # We've seen the last of the translatable strings.  Record the
            # line number of the first line of the strings and update the list
            # of messages seen.  Reset state for the next batch.  If there
            # were no strings inside _(), then just ignore this entry.
            if self.__data:
                self.__addentry(''.join(self.__data))
            self.__state = self.__waiting
        elif ttype == tokenize.STRING:
            self.__data.append(eval(tstring))
        elif ttype not in [tokenize.COMMENT, token.INDENT, token.DEDENT,
                           token.NEWLINE, tokenize.NL]:
            # warn if we see anything else than STRING or whitespace
            print >> sys.stderr, (
                '*** %(file)s:%(lineno)s: Seen unexpected token "%(token)s"'
                ) % {
                'token': tstring,
                'file': self.__curfile,
                'lineno': self.__lineno
                }
            self.__state = self.__waiting

    def __addentry(self, msg, lineno=None, isdocstring=0):
       if lineno is None:
            lineno = self.__lineno
        if not msg in self.__excluded:
            entry = (self.__curfile, lineno)
            self.__messages.setdefault(msg, {})[entry] = isdocstring

    def process(self, filename):
        self.__curfile = filename
        self.__freshmodule = 1

        fp = open(filename)
        try:
            try:
                tokenize.tokenize(fp.readline, self)
            except tokenize.TokenError, e:
                print >> sys.stderr, '%s: %s, line %d, column %d' % (
                    e[0], filename, e[1][0], e[1][1])
        finally:
            fp.close()
        return

    def get_messages(self):
        messages = []

        # Sort the entries.  First sort each particular entry's keys, then
        # sort all the entries by their first item.
        reverse = {}
        for k, v in self.__messages.items():
            keys = v.keys()
            keys.sort()
            reverse.setdefault(tuple(keys), []).append((k, v))
        rkeys = reverse.keys()
        rkeys.sort()
        for rkey in rkeys:
            rentries = reverse[rkey]
            rentries.sort()
            messages.extend(rentries)
        return messages

:: Command execute ::

Enter:
 
Select:
 

:: Search ::
  - regexp 

:: Upload ::
 
[ Read-Only ]

:: Make Dir ::
 
[ Read-Only ]
:: Make File ::
 
[ Read-Only ]

:: Go Dir ::
 
:: Go File ::
 

--[ c99shell v. 1.0 pre-release build #16 powered by Captain Crunch Security Team | http://ccteam.ru | Generation time: 0.0043 ]--