This file is indexed.

/usr/lib/python3/dist-packages/pytest_benchmark-3.0.0.egg-info/PKG-INFO is in python3-pytest-benchmark 3.0.0-1.

This file is owned by root:root, with mode 0o644.

The actual contents of the file can be viewed below.

  1
  2
  3
  4
  5
  6
  7
  8
  9
 10
 11
 12
 13
 14
 15
 16
 17
 18
 19
 20
 21
 22
 23
 24
 25
 26
 27
 28
 29
 30
 31
 32
 33
 34
 35
 36
 37
 38
 39
 40
 41
 42
 43
 44
 45
 46
 47
 48
 49
 50
 51
 52
 53
 54
 55
 56
 57
 58
 59
 60
 61
 62
 63
 64
 65
 66
 67
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
Metadata-Version: 1.1
Name: pytest-benchmark
Version: 3.0.0
Summary: A ``py.test`` fixture for benchmarking code.
Home-page: https://github.com/ionelmc/pytest-benchmark
Author: Ionel Cristian Mărieș
Author-email: contact@ionelmc.ro
License: BSD
Description: ================
        pytest-benchmark
        ================
        
        .. list-table::
            :stub-columns: 1
        
            * - docs
              - |docs|
            * - tests
              - | |travis| |appveyor| |requires| |coveralls| |codecov|
                | |scrutinizer| |codacy| |codeclimate|
            * - package
              - |version| |downloads| |wheel| |supported-versions| |supported-implementations|
        
        .. |docs| image:: https://readthedocs.org/projects/pytest-benchmark/badge/?style=flat
            :target: https://readthedocs.org/projects/pytest-benchmark
            :alt: Documentation Status
        
        .. |travis| image:: https://travis-ci.org/ionelmc/pytest-benchmark.svg?branch=master
            :alt: Travis-CI Build Status
            :target: https://travis-ci.org/ionelmc/pytest-benchmark
        
        .. |appveyor| image:: https://ci.appveyor.com/api/projects/status/github/ionelmc/pytest-benchmark?branch=master&svg=true
            :alt: AppVeyor Build Status
            :target: https://ci.appveyor.com/project/ionelmc/pytest-benchmark
        
        .. |requires| image:: https://requires.io/github/ionelmc/pytest-benchmark/requirements.svg?branch=master
            :alt: Requirements Status
            :target: https://requires.io/github/ionelmc/pytest-benchmark/requirements/?branch=master
        
        .. |coveralls| image:: https://coveralls.io/repos/ionelmc/pytest-benchmark/badge.svg?branch=master&service=github
            :alt: Coverage Status
            :target: https://coveralls.io/r/ionelmc/pytest-benchmark
        
        .. |codecov| image:: https://codecov.io/github/ionelmc/pytest-benchmark/coverage.svg?branch=master
            :alt: Coverage Status
            :target: https://codecov.io/github/ionelmc/pytest-benchmark
        
        .. |landscape| image:: https://landscape.io/github/ionelmc/pytest-benchmark/master/landscape.svg?style=flat
            :target: https://landscape.io/github/ionelmc/pytest-benchmark/master
            :alt: Code Quality Status
        
        .. |codacy| image:: https://img.shields.io/codacy/80e2960677c24d5083a802dd57df17dc.svg?style=flat
            :target: https://www.codacy.com/app/ionelmc/pytest-benchmark
            :alt: Codacy Code Quality Status
        
        .. |codeclimate| image:: https://codeclimate.com/github/ionelmc/pytest-benchmark/badges/gpa.svg
           :target: https://codeclimate.com/github/ionelmc/pytest-benchmark
           :alt: CodeClimate Quality Status
        
        .. |version| image:: https://img.shields.io/pypi/v/pytest-benchmark.svg?style=flat
            :alt: PyPI Package latest release
            :target: https://pypi.python.org/pypi/pytest-benchmark
        
        .. |downloads| image:: https://img.shields.io/pypi/dm/pytest-benchmark.svg?style=flat
            :alt: PyPI Package monthly downloads
            :target: https://pypi.python.org/pypi/pytest-benchmark
        
        .. |wheel| image:: https://img.shields.io/pypi/wheel/pytest-benchmark.svg?style=flat
            :alt: PyPI Wheel
            :target: https://pypi.python.org/pypi/pytest-benchmark
        
        .. |supported-versions| image:: https://img.shields.io/pypi/pyversions/pytest-benchmark.svg?style=flat
            :alt: Supported versions
            :target: https://pypi.python.org/pypi/pytest-benchmark
        
        .. |supported-implementations| image:: https://img.shields.io/pypi/implementation/pytest-benchmark.svg?style=flat
            :alt: Supported implementations
            :target: https://pypi.python.org/pypi/pytest-benchmark
        
        .. |scrutinizer| image:: https://img.shields.io/scrutinizer/g/ionelmc/pytest-benchmark/master.svg?style=flat
            :alt: Scrutinizer Status
            :target: https://scrutinizer-ci.com/g/ionelmc/pytest-benchmark/
        
        A ``py.test`` fixture for benchmarking code. It will group the tests into rounds that are calibrated to the chosen
        timer. See calibration_ and FAQ_.
        
        * Free software: BSD license
        
        Installation
        ============
        
        ::
        
            pip install pytest-benchmark
        
        Documentation
        =============
        
        Available at: `pytest-benchmark.readthedocs.org <http://pytest-benchmark.readthedocs.org/en/stable/>`_.
        
        Examples
        ========
        
        This plugin provides a `benchmark` fixture. This fixture is a callable object that will benchmark any function passed
        to it.
        
        Example:
        
        .. code-block:: python
        
            def something(duration=0.000001):
                """
                Function that needs some serious benchmarking.
                """
                time.sleep(duration)
                # You may return anything you want, like the result of a computation
                return 123
        
            def test_my_stuff(benchmark):
                # benchmark something
                result = benchmark(something)
        
                # Extra code, to verify that the run completed correctly.
                # Sometimes you may want to check the result, fast functions
                # are no good if they return incorrect results :-)
                assert result == 123
        
        You can also pass extra arguments:
        
        .. code-block:: python
        
            def test_my_stuff(benchmark):
                benchmark(time.sleep, 0.02)
        
        Or even keyword arguments:
        
        .. code-block:: python
        
            def test_my_stuff(benchmark):
                benchmark(time.sleep, duration=0.02)
        
        Another pattern seen in the wild, that is not recommended for micro-benchmarks (very fast code) but may be convenient:
        
        .. code-block:: python
        
            def test_my_stuff(benchmark):
                @benchmark
                def something():  # unnecessary function call
                    time.sleep(0.000001)
        
        A better way is to just benchmark the final function:
        
        .. code-block:: python
        
            def test_my_stuff(benchmark):
                benchmark(time.sleep, 0.000001)  # way more accurate results!
        
        If you need to do fine control over how the benchmark is run (like a `setup` function, exact control of `iterations` and
        `rounds`) there's a special mode - pedantic_:
        
        .. code-block:: python
        
            def my_special_setup():
                ...
        
            def test_with_setup(benchmark):
                benchmark.pedantic(something, setup=my_special_setup, args=(1, 2, 3), kwargs={'foo': 'bar'}, iterations=10, rounds=100)
        
        Screenshots
        -----------
        
        Normal run:
        
        .. image:: https://github.com/ionelmc/pytest-benchmark/raw/master/docs/screenshot.png
            :alt: Screenshot of py.test summary
        
        Compare mode (``--benchmark-compare``):
        
        .. image:: https://github.com/ionelmc/pytest-benchmark/raw/master/docs/screenshot-compare.png
            :alt: Screenshot of py.test summary in compare mode
        
        Histogram (``--benchmark-histogram``):
        
        .. image:: https://cdn.rawgit.com/ionelmc/pytest-benchmark/94860cc8f47aed7ba4f9c7e1380c2195342613f6/docs/sample-tests_test_normal.py_test_xfast_parametrized%5B0%5D.svg
            :alt: Histogram sample
        
        ..
        
            Also, it has `nice tooltips <https://cdn.rawgit.com/ionelmc/pytest-benchmark/master/docs/sample.svg>`_.
        
        Development
        ===========
        
        To run the all tests run::
        
            tox
        
        Credits
        =======
        
        * Timing code and ideas taken from: https://bitbucket.org/haypo/misc/src/tip/python/benchmark.py
        
        .. _FAQ: http://pytest-benchmark.readthedocs.org/en/latest/faq.html
        .. _calibration: http://pytest-benchmark.readthedocs.org/en/latest/features.html#calibration
        .. _pedantic: http://pytest-benchmark.readthedocs.org/en/latest/pedantic.html
        
        
        
        
        
        
        Changelog
        =========
        
        3.0.0 (2015-08-11)
        ------------------
        
        * Improved ``--help`` text for ``--benchmark-histogram``, ``--benchmark-save`` and ``--benchmark-autosave``.
        * Benchmarks that raised exceptions during test now have special highlighting in result table (red background).
        * Benchmarks that raised exceptions are not included in the saved data anymore (you can still get the old behavior back
          by implementing ``pytest_benchmark_generate_json`` in your ``conftest.py``).
        * The plugin will use pytest's warning system for warnings. There are 2 categories: ``WBENCHMARK-C`` (compare mode
          issues) and ``WBENCHMARK-U`` (usage issues).
        * The red warnings are only shown if ``--benchmark-verbose`` is used. They still will be always be shown in the
          pytest-warnings section.
        * Using the benchmark fixture more than one time is disallowed (will raise exception).
        * Not using the benchmark fixutre (but requiring it) will issue a warning (``WBENCHMARK-U1``).
        
        3.0.0rc1 (2015-10-25)
        ---------------------
        
        * Changed ``--benchmark-warmup`` to take optional value and automatically activate on PyPy (default value is ``auto``).
          *MAY BE BACKWARDS INCOMPATIBLE*
        * Removed the version check in compare mode (previously there was a warning if current version is lower than what's in
          the file).
        
        3.0.0b3 (2015-10-22)
        ---------------------
        
        * Changed how comparison is displayed in the result table. Now previous runs are shown as normal runs and names get a
          special suffix indicating the origin. Eg: "test_foobar (NOW)" or "test_foobar (0123)".
        * Fixed sorting in the result table. Now rows are sorted by the sort column, and then by name.
        * Show the plugin version in the header section.
        * Moved the display of default options in the header section.
        
        3.0.0b2 (2015-10-17)
        ---------------------
        
        * Add a ``--benchmark-disable`` option. It's automatically activated when xdist is on
        * When xdist is on or `statistics` can't be imported then ``--benchmark-disable`` is automatically activated (instead
          of ``--benchmark-skip``). *BACKWARDS INCOMPATIBLE*
        * Replace the deprecated ``__multicall__`` with the new hookwrapper system.
        * Improved description for ``--benchmark-max-time``.
        
        3.0.0b1 (2015-10-13)
        --------------------
        
        * Tests are sorted alphabetically in the results table.
        * Failing to import `statistics` doesn't create hard failures anymore. Benchmarks are automatically skipped if import
          failure occurs. This would happen on Python 3.2 (or earlier Python 3).
        
        3.0.0a4 (2015-10-08)
        --------------------
        
        * Changed how failures to get commit info are handled: now they are soft failures. Previously it made the whole
          test suite fail, just because you didn't have ``git/hg`` installed.
        
        3.0.0a3 (2015-10-02)
        --------------------
        
        * Added progress indication when computing stats.
        
        3.0.0a2 (2015-09-30)
        --------------------
        
        * Fixed accidental output capturing caused by capturemanager misuse.
        
        3.0.0a1 (2015-09-13)
        --------------------
        
        * Added JSON report saving (the ``--benchmark-json`` command line arguments).
        * Added benchmark data storage(the ``--benchmark-save`` and ``--benchmark-autosave`` command line arguments).
        * Added comparison to previous runs (the ``--benchmark-compare`` command line argument).
        * Added performance regression checks (the ``--benchmark-compare-fail`` command line argument).
        * Added possibility to group by various parts of test name (the `--benchmark-compare-group-by`` command line argument).
        * Added historical plotting (the ``--benchmark-histogram`` command line argument).
        * Added option to fine tune the calibration (the ``--benchmark-calibration-precision`` command line argument and
          ``calibration_precision`` marker option).
        
        * Changed ``benchmark_weave`` to no longer be a context manager. Cleanup is performed automatically. *BACKWARDS
          INCOMPATIBLE*
        * Added ``benchmark.weave`` method (alternative to ``benchmark_weave`` fixture).
        
        * Added new hooks to allow customization:
        
          * ``pytest_benchmark_generate_machine_info(config)``
          * ``pytest_benchmark_update_machine_info(config, info)``
          * ``pytest_benchmark_generate_commit_info(config)``
          * ``pytest_benchmark_update_commit_info(config, info)``
          * ``pytest_benchmark_group_stats(config, benchmarks, group_by)``
          * ``pytest_benchmark_generate_json(config, benchmarks, include_data)``
          * ``pytest_benchmark_update_json(config, benchmarks, output_json)``
          * ``pytest_benchmark_compare_machine_info(config, benchmarksession, machine_info, compared_benchmark)``
        
        * Changed the timing code to:
        
          * Tracers are automatically disabled when running the test function (like coverage tracers).
          * Fixed an issue with calibration code getting stuck.
        
        * Added `pedantic mode` via ``benchmark.pedantic()``. This mode disables calibration and allows a setup function.
        
        
        2.5.0 (2015-06-20)
        ------------------
        
        * Improved test suite a bit (not using `cram` anymore).
        * Improved help text on the ``--benchmark-warmup`` option.
        * Made ``warmup_iterations`` available as a marker argument (eg: ``@pytest.mark.benchmark(warmup_iterations=1234)``).
        * Fixed ``--benchmark-verbose``'s printouts to work properly with output capturing.
        * Changed how warmup iterations are computed (now number of total iterations is used, instead of just the rounds).
        * Fixed a bug where calibration would run forever.
        * Disabled red/green coloring (it was kinda random) when there's a single test in the results table.
        
        2.4.1 (2015-03-16)
        ------------------
        
        * Fix regression, plugin was raising ``ValueError: no option named 'dist'`` when xdist wasn't installed.
        
        2.4.0 (2015-03-12)
        ------------------
        
        * Add a ``benchmark_weave`` experimental fixture.
        * Fix internal failures when `xdist` plugin is active.
        * Automatically disable benchmarks if `xdist` is active.
        
        2.3.0 (2014-12-27)
        ------------------
        
        * Moved the warmup in the calibration phase. Solves issues with benchmarking on PyPy.
        
          Added a ``--benchmark-warmup-iterations`` option to fine-tune that.
        
        2.2.0 (2014-12-26)
        ------------------
        
        * Make the default rounds smaller (so that variance is more accurate).
        * Show the defaults in the ``--help`` section.
        
        2.1.0 (2014-12-20)
        ------------------
        
        * Simplify the calibration code so that the round is smaller.
        * Add diagnostic output for calibration code (``--benchmark-verbose``).
        
        2.0.0 (2014-12-19)
        ------------------
        
        * Replace the context-manager based API with a simple callback interface. *BACKWARDS INCOMPATIBLE*
        * Implement timer calibration for precise measurements.
        
        1.0.0 (2014-12-15)
        ------------------
        
        * Use a precise default timer for PyPy.
        
        ? (?)
        -----
        
        * Readme and styling fixes (contributed by Marc Abramowitz)
        * Lots of wild changes.
        
Keywords: pytest,benchmark
Platform: UNKNOWN
Classifier: Development Status :: 5 - Production/Stable
Classifier: Intended Audience :: Developers
Classifier: License :: OSI Approved :: BSD License
Classifier: Operating System :: Unix
Classifier: Operating System :: POSIX
Classifier: Operating System :: Microsoft :: Windows
Classifier: Programming Language :: Python
Classifier: Programming Language :: Python :: 2.6
Classifier: Programming Language :: Python :: 2.7
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.3
Classifier: Programming Language :: Python :: 3.4
Classifier: Programming Language :: Python :: 3.5
Classifier: Programming Language :: Python :: Implementation :: CPython
Classifier: Programming Language :: Python :: Implementation :: PyPy
Classifier: Topic :: Utilities