Files
@ 190cb30841de
Branch filter:
Location: kallithea/docs/usage/performance.rst - annotation
190cb30841de
3.7 KiB
text/prs.fallenstein.rst
branches: fix performance of branch selectors with many branches - only show the first 200 results
The way we use select2, it will cause browser performance problems when a
select list contains thousands of entries. The primary bottleneck is the DOM
creation, secondarily for the query to filter through the entries and decide
what to show. We thus primarily have to limit how many entries we put in the
drop-down, secondarily limit the iteration over data.
One tricky case is where the user specifies a short but full branch name (like
'trunk') but many other branches contains the same string (not necessarily at
the beginning, like 'for-trunk-next-week') which come before the perfect match
in the branch list. It is thus not a solution to just stop searching when a
fixed amount of matches have been found.
Instead, we limit the amount of ordinary query matches, but always show all
prefix matches. We thus always have to iterate through all entries, but we
start using the (presumably) cheaper prefix search when the limit has been
reached.
There is no filtering initially when there is no query term, so that case has
to be handled specially.
Upstream select2 is now at 4.x. Upgrading is not trivial, and getting this
fixed properly upstream is not a short term solution. Instead, we customize our
copy. The benefit from this patch is bigger than the overhead of "maintaining"
it locally.
The way we use select2, it will cause browser performance problems when a
select list contains thousands of entries. The primary bottleneck is the DOM
creation, secondarily for the query to filter through the entries and decide
what to show. We thus primarily have to limit how many entries we put in the
drop-down, secondarily limit the iteration over data.
One tricky case is where the user specifies a short but full branch name (like
'trunk') but many other branches contains the same string (not necessarily at
the beginning, like 'for-trunk-next-week') which come before the perfect match
in the branch list. It is thus not a solution to just stop searching when a
fixed amount of matches have been found.
Instead, we limit the amount of ordinary query matches, but always show all
prefix matches. We thus always have to iterate through all entries, but we
start using the (presumably) cheaper prefix search when the limit has been
reached.
There is no filtering initially when there is no query term, so that case has
to be handled specially.
Upstream select2 is now at 4.x. Upgrading is not trivial, and getting this
fixed properly upstream is not a short term solution. Instead, we customize our
copy. The benefit from this patch is bigger than the overhead of "maintaining"
it locally.
fa88997aa421 fa88997aa421 fa88997aa421 22a3fa3c4254 fa88997aa421 fa88997aa421 4e6dfdb3fa01 4e6dfdb3fa01 4e6dfdb3fa01 5d12768a0aa1 5d12768a0aa1 e73a69cb98dc fa88997aa421 6afa528ee30e 6afa528ee30e 6afa528ee30e 6afa528ee30e fa88997aa421 6257de126ec7 4610a39d3be9 fa88997aa421 6afa528ee30e 8b8edfc25856 5ae8e644aa88 5ae8e644aa88 5ae8e644aa88 778f7ae3b6eb 778f7ae3b6eb 8b8edfc25856 6afa528ee30e fa88997aa421 5ae8e644aa88 5ae8e644aa88 4cd84f4f28fb 4cd84f4f28fb d79f3505549e 4e6dfdb3fa01 4e6dfdb3fa01 4e6dfdb3fa01 4e6dfdb3fa01 4e6dfdb3fa01 4e6dfdb3fa01 4e6dfdb3fa01 4e6dfdb3fa01 4e6dfdb3fa01 4e6dfdb3fa01 4cd84f4f28fb 4cd84f4f28fb 778f7ae3b6eb 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df 61954577a0df fbbe80e3322b 778f7ae3b6eb | .. _performance:
================================
Optimizing Kallithea performance
================================
When serving a large amount of big repositories, Kallithea can start
performing slower than expected. Because of the demanding nature of handling large
amounts of data from version control systems, here are some tips on how to get
the best performance.
Follow these few steps to improve performance of Kallithea system.
1. Kallithea is often I/O bound, and hence a fast disk (SSD/SAN) is
usually more important than a fast CPU.
2. Increase cache
Tweak beaker cache settings in the ini file. The actual effect of that
is questionable.
3. Switch from SQLite to PostgreSQL or MySQL
SQLite is a good option when having a small load on the system. But due to
locking issues with SQLite, it is not recommended to use it for larger
deployments. Switching to MySQL or PostgreSQL will result in an immediate
performance increase. A tool like SQLAlchemyGrate_ can be used for
migrating to another database platform.
4. Scale Kallithea horizontally
Scaling horizontally can give huge performance benefits when dealing with
large amounts of traffic (many users, CI servers, etc.). Kallithea can be
scaled horizontally on one (recommended) or multiple machines. In order
to scale horizontally you need to do the following:
- Each instance's ``data`` storage needs to be configured to be stored on a
shared disk storage, preferably together with repositories. This ``data``
dir contains template caches, sessions, whoosh index and is used for
task locking (so it is safe across multiple instances). Set the
``cache_dir``, ``index_dir``, ``beaker.cache.data_dir``, ``beaker.cache.lock_dir``
variables in each .ini file to a shared location across Kallithea instances
- If celery is used each instance should run a separate Celery instance, but
the message broker should be common to all of them (e.g., one
shared RabbitMQ server)
- Load balance using round robin or IP hash, recommended is writing LB rules
that will separate regular user traffic from automated processes like CI
servers or build bots.
5. Serve static files directly from the web server
With the default ``static_files`` ini setting, the Kallithea WSGI application
will take care of serving the static files found in ``kallithea/public`` from
the root of the application URL. While doing that, it will currently also
apply buffering and compression of all the responses it is serving.
The actual serving of the static files is unlikely to be a problem in a
Kallithea setup. The buffering of responses is more likely to be a problem;
large responses (clones or pulls) will have to be fully processed and spooled
to disk or memory before the client will see any response.
To serve static files from the web server, use something like this Apache config
snippet::
Alias /images/ /srv/kallithea/kallithea/kallithea/public/images/
Alias /css/ /srv/kallithea/kallithea/kallithea/public/css/
Alias /js/ /srv/kallithea/kallithea/kallithea/public/js/
Alias /codemirror/ /srv/kallithea/kallithea/kallithea/public/codemirror/
Alias /fontello/ /srv/kallithea/kallithea/kallithea/public/fontello/
Then disable serving of static files in the ``.ini`` ``app:main`` section::
static_files = false
If using Kallithea installed as a package, you should be able to find the files
under site-packages/kallithea, either in your Python installation or in your
virtualenv. When upgrading, make sure to update the web server configuration
too if necessary.
.. _SQLAlchemyGrate: https://github.com/shazow/sqlalchemygrate
|