Files @ 0eceb478c720
Branch filter:

Location: kallithea/rhodecode/lib/celerylib/tasks.py - annotation

Marcin Kuzminski
fixed issue #165 trending source files are now stored in cache as ext only, and translated to description only when displaying, so future changes of mappings will take affect right away.
Also changes the way how map is generating, ie. autogenerate from pygments lexers, and than complete with user defined mappings.
04c9bb9ca6d6
04c9bb9ca6d6
04c9bb9ca6d6
3a7f5b1a19dd
04c9bb9ca6d6
04c9bb9ca6d6
04c9bb9ca6d6
6832ef664673
04c9bb9ca6d6
04c9bb9ca6d6
6832ef664673
04c9bb9ca6d6
04c9bb9ca6d6
a671db5bdd58
a671db5bdd58
a671db5bdd58
a671db5bdd58
6832ef664673
04c9bb9ca6d6
04c9bb9ca6d6
04c9bb9ca6d6
04c9bb9ca6d6
6832ef664673
04c9bb9ca6d6
a671db5bdd58
1e757ac98988
03676d39dd0a
cb0d9ce6ac5c
cb0d9ce6ac5c
3a7f5b1a19dd
3a7f5b1a19dd
cb0d9ce6ac5c
1e757ac98988
0eceb478c720
0eceb478c720
f6c613fba757
f6c613fba757
1e757ac98988
f6c613fba757
f6c613fba757
1e757ac98988
1e757ac98988
277427ac29a9
f6c613fba757
f6c613fba757
f6c613fba757
f6c613fba757
05528ad948c4
f6c613fba757
f6c613fba757
cb0d9ce6ac5c
aac24db58ce8
aac24db58ce8
03676d39dd0a
ba7e24cd4786
ba7e24cd4786
ba7e24cd4786
ba7e24cd4786
ba7e24cd4786
1e757ac98988
1e757ac98988
1e757ac98988
f6c613fba757
f6c613fba757
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
1e757ac98988
f6c613fba757
f6c613fba757
f6c613fba757
f6c613fba757
1e757ac98988
1e757ac98988
0eceb478c720
cb0d9ce6ac5c
cb0d9ce6ac5c
cb0d9ce6ac5c
cb0d9ce6ac5c
cb0d9ce6ac5c
0eceb478c720
3a7f5b1a19dd
1e757ac98988
1e757ac98988
3a7f5b1a19dd
1e757ac98988
f6c613fba757
373ee7031003
aac24db58ce8
aac24db58ce8
1e757ac98988
0eceb478c720
3a7f5b1a19dd
1e757ac98988
1e757ac98988
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
1e757ac98988
52da7cba88a6
52da7cba88a6
0eceb478c720
7e536d1af60d
0eceb478c720
1e757ac98988
cb0d9ce6ac5c
05528ad948c4
fd63782c4426
1e757ac98988
1e757ac98988
db71228a4c23
1e757ac98988
1e757ac98988
1e757ac98988
7e536d1af60d
1e757ac98988
7e536d1af60d
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
61c5f0ab24b9
61c5f0ab24b9
1e757ac98988
61c5f0ab24b9
61c5f0ab24b9
1e757ac98988
1e757ac98988
7e536d1af60d
61c5f0ab24b9
6832ef664673
52da7cba88a6
1e757ac98988
7e536d1af60d
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
0eceb478c720
7e536d1af60d
1e757ac98988
a9e72f7148c8
a9e72f7148c8
317ba4c84160
db71228a4c23
db71228a4c23
0eceb478c720
b92d9a0321e0
b92d9a0321e0
b92d9a0321e0
0eceb478c720
1e757ac98988
0eceb478c720
0eceb478c720
1e757ac98988
1e757ac98988
1e757ac98988
7e536d1af60d
1e757ac98988
7e536d1af60d
0eceb478c720
7e536d1af60d
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
7e536d1af60d
1e757ac98988
1e757ac98988
7e536d1af60d
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
1e757ac98988
0eceb478c720
0eceb478c720
7e536d1af60d
1e757ac98988
1e757ac98988
0eceb478c720
0eceb478c720
0eceb478c720
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
0eceb478c720
7e536d1af60d
7e536d1af60d
1a77a0c350f5
0eceb478c720
1e757ac98988
1e757ac98988
1e757ac98988
7e536d1af60d
b92d9a0321e0
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
0eceb478c720
1e757ac98988
1e757ac98988
1e757ac98988
0eceb478c720
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
7e536d1af60d
7e536d1af60d
25116ab9f0bc
1e757ac98988
7e536d1af60d
1e757ac98988
db71228a4c23
db71228a4c23
1e757ac98988
7e536d1af60d
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
7e536d1af60d
1e757ac98988
1e757ac98988
0eceb478c720
3a7f5b1a19dd
1e757ac98988
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
1e757ac98988
1e757ac98988
7e536d1af60d
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
716911af91e1
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
7e536d1af60d
1e757ac98988
1e757ac98988
1e757ac98988
7e536d1af60d
1e757ac98988
f99075170eb4
f99075170eb4
1e757ac98988
7e536d1af60d
1e757ac98988
1e757ac98988
1e757ac98988
f6c613fba757
1e757ac98988
1e757ac98988
0eceb478c720
3a7f5b1a19dd
1e757ac98988
ecc566f8b69f
ecc566f8b69f
6832ef664673
6832ef664673
ecc566f8b69f
ecc566f8b69f
ecc566f8b69f
ecc566f8b69f
ecc566f8b69f
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
f6c613fba757
ecc566f8b69f
ecc566f8b69f
ecc566f8b69f
ecc566f8b69f
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
ecc566f8b69f
ecc566f8b69f
f6dca275c5a8
7e536d1af60d
1e757ac98988
1e757ac98988
f6dca275c5a8
7e536d1af60d
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
1e757ac98988
0eceb478c720
3a7f5b1a19dd
1e757ac98988
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
3a7f5b1a19dd
7e536d1af60d
758f64f3fbda
3a7f5b1a19dd
89b9037d68b7
a9e72f7148c8
758f64f3fbda
cb0d9ce6ac5c
758f64f3fbda
1e757ac98988
758f64f3fbda
7e536d1af60d
758f64f3fbda
758f64f3fbda
758f64f3fbda
7e536d1af60d
0eceb478c720
1e757ac98988
cb0d9ce6ac5c
05528ad948c4
fd63782c4426
95a502d94860
1e757ac98988
a9e72f7148c8
a9e72f7148c8
a9e72f7148c8
0eceb478c720
9fec2207f91c
0eceb478c720
0eceb478c720
1e757ac98988
0eceb478c720
7e536d1af60d
a9e72f7148c8
a9e72f7148c8
1e757ac98988
# -*- coding: utf-8 -*-
"""
    rhodecode.lib.celerylib.tasks
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

    RhodeCode task modules, containing all task that suppose to be run
    by celery daemon

    :created_on: Oct 6, 2010
    :author: marcink
    :copyright: (C) 2009-2011 Marcin Kuzminski <marcin@python-works.com>
    :license: GPLv3, see COPYING for more details.
"""
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program.  If not, see <http://www.gnu.org/licenses/>.
from celery.decorators import task

import os
import traceback
import logging

from time import mktime
from operator import itemgetter
from pygments import lexers
from string import lower

from pylons import config
from pylons.i18n.translation import _

from rhodecode.lib.celerylib import run_task, locked_task, str2bool
from rhodecode.lib.helpers import person
from rhodecode.lib.smtp_mailer import SmtpMailer
from rhodecode.lib.utils import OrderedDict, add_cache
from rhodecode.model import init_model
from rhodecode.model import meta
from rhodecode.model.db import RhodeCodeUi

from vcs.backends import get_repo

from sqlalchemy import engine_from_config

add_cache(config)

try:
    import json
except ImportError:
    #python 2.5 compatibility
    import simplejson as json

__all__ = ['whoosh_index', 'get_commits_stats',
           'reset_user_password', 'send_email']

CELERY_ON = str2bool(config['app_conf'].get('use_celery'))

LANGUAGES_EXTENSIONS_MAP = {}


def __clean(s):

    s = s.lstrip('*')
    s = s.lstrip('.')

    if s.find('[') != -1:
        exts = []
        start, stop = s.find('['), s.find(']')

        for suffix in s[start + 1:stop]:
            exts.append(s[:s.find('[')] + suffix)
        return map(lower, exts)
    else:
        return map(lower, [s])

for lx, t in sorted(lexers.LEXERS.items()):
    m = map(__clean, t[-2])
    if m:
        m = reduce(lambda x, y: x + y, m)
        for ext in m:
            desc = lx.replace('Lexer', '')
            if ext in LANGUAGES_EXTENSIONS_MAP:
                if desc not in LANGUAGES_EXTENSIONS_MAP[ext]:
                    LANGUAGES_EXTENSIONS_MAP[ext].append(desc)
            else:
                LANGUAGES_EXTENSIONS_MAP[ext] = [desc]

#Additional mappings that are not present in the pygments lexers
# NOTE: that this will overide any mappings in LANGUAGES_EXTENSIONS_MAP
ADDITIONAL_MAPPINGS = {'xaml': 'XAML'}

LANGUAGES_EXTENSIONS_MAP.update(ADDITIONAL_MAPPINGS)


def get_session():
    if CELERY_ON:
        engine = engine_from_config(config, 'sqlalchemy.db1.')
        init_model(engine)
    sa = meta.Session()
    return sa


def get_repos_path():
    sa = get_session()
    q = sa.query(RhodeCodeUi).filter(RhodeCodeUi.ui_key == '/').one()
    return q.ui_value


@task(ignore_result=True)
@locked_task
def whoosh_index(repo_location, full_index):
    #log = whoosh_index.get_logger()
    from rhodecode.lib.indexers.daemon import WhooshIndexingDaemon
    index_location = config['index_dir']
    WhooshIndexingDaemon(index_location=index_location,
                         repo_location=repo_location, sa=get_session())\
                         .run(full_index=full_index)


@task(ignore_result=True)
@locked_task
def get_commits_stats(repo_name, ts_min_y, ts_max_y):
    try:
        log = get_commits_stats.get_logger()
    except:
        log = logging.getLogger(__name__)

    from rhodecode.model.db import Statistics, Repository

    #for js data compatibilty
    akc = lambda k: person(k).replace('"', "")

    co_day_auth_aggr = {}
    commits_by_day_aggregate = {}
    repos_path = get_repos_path()
    p = os.path.join(repos_path, repo_name)
    repo = get_repo(p)

    skip_date_limit = True
    parse_limit = int(config['app_conf'].get('commit_parse_limit'))
    last_rev = 0
    last_cs = None
    timegetter = itemgetter('time')

    sa = get_session()

    dbrepo = sa.query(Repository)\
        .filter(Repository.repo_name == repo_name).scalar()
    cur_stats = sa.query(Statistics)\
        .filter(Statistics.repository == dbrepo).scalar()

    if cur_stats is not None:
        last_rev = cur_stats.stat_on_revision

    #return if repo is empty
    if not repo.revisions:
        return True

    if last_rev == repo.get_changeset().revision and len(repo.revisions) > 1:
        #pass silently without any work if we're not on first revision or
        #current state of parsing revision(from db marker) is the last revision
        return True

    if cur_stats:
        commits_by_day_aggregate = OrderedDict(
                                       json.loads(
                                        cur_stats.commit_activity_combined))
        co_day_auth_aggr = json.loads(cur_stats.commit_activity)

    log.debug('starting parsing %s', parse_limit)
    lmktime = mktime

    last_rev = last_rev + 1 if last_rev > 0 else last_rev

    for cs in repo[last_rev:last_rev + parse_limit]:
        last_cs = cs  # remember last parsed changeset
        k = lmktime([cs.date.timetuple()[0], cs.date.timetuple()[1],
                      cs.date.timetuple()[2], 0, 0, 0, 0, 0, 0])

        if akc(cs.author) in co_day_auth_aggr:
            try:
                l = [timegetter(x) for x in
                     co_day_auth_aggr[akc(cs.author)]['data']]
                time_pos = l.index(k)
            except ValueError:
                time_pos = False

            if time_pos >= 0 and time_pos is not False:

                datadict = co_day_auth_aggr[akc(cs.author)]['data'][time_pos]

                datadict["commits"] += 1
                datadict["added"] += len(cs.added)
                datadict["changed"] += len(cs.changed)
                datadict["removed"] += len(cs.removed)

            else:
                if k >= ts_min_y and k <= ts_max_y or skip_date_limit:

                    datadict = {"time": k,
                                "commits": 1,
                                "added": len(cs.added),
                                "changed": len(cs.changed),
                                "removed": len(cs.removed),
                               }
                    co_day_auth_aggr[akc(cs.author)]['data']\
                        .append(datadict)

        else:
            if k >= ts_min_y and k <= ts_max_y or skip_date_limit:
                co_day_auth_aggr[akc(cs.author)] = {
                                    "label": akc(cs.author),
                                    "data": [{"time":k,
                                             "commits":1,
                                             "added":len(cs.added),
                                             "changed":len(cs.changed),
                                             "removed":len(cs.removed),
                                             }],
                                    "schema": ["commits"],
                                    }

        #gather all data by day
        if k in commits_by_day_aggregate:
            commits_by_day_aggregate[k] += 1
        else:
            commits_by_day_aggregate[k] = 1

    overview_data = sorted(commits_by_day_aggregate.items(), key=itemgetter(0))
    if not co_day_auth_aggr:
        co_day_auth_aggr[akc(repo.contact)] = {
            "label": akc(repo.contact),
            "data": [0, 1],
            "schema": ["commits"],
        }

    stats = cur_stats if cur_stats else Statistics()
    stats.commit_activity = json.dumps(co_day_auth_aggr)
    stats.commit_activity_combined = json.dumps(overview_data)

    log.debug('last revison %s', last_rev)
    leftovers = len(repo.revisions[last_rev:])
    log.debug('revisions to parse %s', leftovers)

    if last_rev == 0 or leftovers < parse_limit:
        log.debug('getting code trending stats')
        stats.languages = json.dumps(__get_codes_stats(repo_name))

    try:
        stats.repository = dbrepo
        stats.stat_on_revision = last_cs.revision if last_cs else 0
        sa.add(stats)
        sa.commit()
    except:
        log.error(traceback.format_exc())
        sa.rollback()
        return False
    if len(repo.revisions) > 1:
        run_task(get_commits_stats, repo_name, ts_min_y, ts_max_y)

    return True


@task(ignore_result=True)
def reset_user_password(user_email):
    try:
        log = reset_user_password.get_logger()
    except:
        log = logging.getLogger(__name__)

    from rhodecode.lib import auth
    from rhodecode.model.db import User

    try:
        try:
            sa = get_session()
            user = sa.query(User).filter(User.email == user_email).scalar()
            new_passwd = auth.PasswordGenerator().gen_password(8,
                             auth.PasswordGenerator.ALPHABETS_BIG_SMALL)
            if user:
                user.password = auth.get_crypt_password(new_passwd)
                user.api_key = auth.generate_api_key(user.username)
                sa.add(user)
                sa.commit()
                log.info('change password for %s', user_email)
            if new_passwd is None:
                raise Exception('unable to generate new password')

        except:
            log.error(traceback.format_exc())
            sa.rollback()

        run_task(send_email, user_email,
                 "Your new rhodecode password",
                 'Your new rhodecode password:%s' % (new_passwd))
        log.info('send new password mail to %s', user_email)

    except:
        log.error('Failed to update user password')
        log.error(traceback.format_exc())

    return True


@task(ignore_result=True)
def send_email(recipients, subject, body):
    """
    Sends an email with defined parameters from the .ini files.


    :param recipients: list of recipients, it this is empty the defined email
        address from field 'email_to' is used instead
    :param subject: subject of the mail
    :param body: body of the mail
    """
    try:
        log = send_email.get_logger()
    except:
        log = logging.getLogger(__name__)

    email_config = config

    if not recipients:
        recipients = [email_config.get('email_to')]

    mail_from = email_config.get('app_email_from')
    user = email_config.get('smtp_username')
    passwd = email_config.get('smtp_password')
    mail_server = email_config.get('smtp_server')
    mail_port = email_config.get('smtp_port')
    tls = str2bool(email_config.get('smtp_use_tls'))
    ssl = str2bool(email_config.get('smtp_use_ssl'))
    debug = str2bool(config.get('debug'))

    try:
        m = SmtpMailer(mail_from, user, passwd, mail_server,
                       mail_port, ssl, tls, debug=debug)
        m.send(recipients, subject, body)
    except:
        log.error('Mail sending failed')
        log.error(traceback.format_exc())
        return False
    return True


@task(ignore_result=True)
def create_repo_fork(form_data, cur_user):
    try:
        log = create_repo_fork.get_logger()
    except:
        log = logging.getLogger(__name__)

    from rhodecode.model.repo import RepoModel
    from vcs import get_backend

    repo_model = RepoModel(get_session())
    repo_model.create(form_data, cur_user, just_db=True, fork=True)
    repo_name = form_data['repo_name']
    repos_path = get_repos_path()
    repo_path = os.path.join(repos_path, repo_name)
    repo_fork_path = os.path.join(repos_path, form_data['fork_name'])
    alias = form_data['repo_type']

    log.info('creating repo fork %s as %s', repo_name, repo_path)
    backend = get_backend(alias)
    backend(str(repo_fork_path), create=True, src_url=str(repo_path))


def __get_codes_stats(repo_name):
    repos_path = get_repos_path()
    p = os.path.join(repos_path, repo_name)
    repo = get_repo(p)
    tip = repo.get_changeset()
    code_stats = {}

    def aggregate(cs):
        for f in cs[2]:
            ext = lower(f.extension)
            if ext in LANGUAGES_EXTENSIONS_MAP.keys() and not f.is_binary:
                if ext in code_stats:
                    code_stats[ext] += 1
                else:
                    code_stats[ext] = 1

    map(aggregate, tip.walk('/'))

    return code_stats or {}